When it comes to enterprise adoption of generative AI, most organizations are in transition. While 88% of customers we've talked to say they are currently running GenAI pilot projects,the majority also say they are too nervous to take those experiments from the test environment to production.
So, what’s causing this disparity? Concerns around cost and risk. In the past, when it came to IT investments, companies could take a “build it, and they will come” mentality. Not anymore. Now, new projects are expected to produce value for the business – and quickly. In the past, board members and investors may have been fine waiting several years for a return on IT investment; now they want to see progress in as little as six months.
Not only are enterprises concerned about the ROI of GenAI development costs, but they also worry that AI systems could spit out bad or inaccurate results (i.e., hallucinate) that might harm their business or expose sensitive or confidential company information. As well, legal departments are now taking a closer look at technology projects than ever before. They want assurances that systems are generating explainable and trustworthy results. Meanwhile, operations teams want to be sure they can control who or what is able to access proprietary information and that their data is used in a compliant manner.
With apologies to Robert Johnson, if you’re standing at the crossroads, don’t let innovations powered by GenAI pass you by. Test environments can only provide so much information, and companies won’t understand and benefit from the true value of AI systems until they are deployed in the real world. However, getting to that point often requires an organizational overhaul.
To move GenAI projects from experimentation to production (and scale them across the enterprise), companies must ensure that the compound AI systems that power these applications are accurate, safe, and governed.
Accurate
There’s an old truism in computer science: “Garbage in, garbage out.” In other words, good AI requires good data. For an AI model to produce accurate and contextually relevant results, it needs quality, relevant data as its input.
Off-the-shelf commercial models can lack the necessary knowledge about a company’s unique operations to produce insights that will deliver enough relevant business impact. These models may misinterpret company jargon or return information that is inaccurate in the context of the business. For example, Databricks employees are known internally as “Bricksters,” a definition that doesn’t show up in queries to public models.