We’re excited to introduce the beta availability of Oracle Cloud Infrastructure (OCI) Generative AI Agents Retrieval-Augmented Generation (RAG) service, your organization’s own customizable solution for conversing with and acting on diverse knowledge bases. The RAG service is the first of a series of AI Agents, with an focus on OpenSearch. Upcoming releases are expected to support a wider range of large language models (LLMs) and provide access to Oracle Database 23c with AI Vector Search and MySQL HeatWave with Vector Store.
We’ve all heard the buzz around ChatGPT and other LLMs that can understand and respond to natural language requests. Now, Oracle brings similar capabilities purpose-built for enterprises, by combining user-friendly conversational interfaces with seamless real-time access to corporate knowledge bases.
For example, OCI Generative AI Agents RAG service allows employees to ask questions and get insights from siloed sources across your company through natural conversations. The RAG service performs a semantic similarity match between the user’s request and the documents in a centralized embeddings vector store. The embeddings vector store can contain data from a variety of corporate systems, such as enterprise resource planning (ERP), Human Capital Management (HCM), Customer Experience (CX), and data lakes, and query multiple data silos simultaneously. Users aren’t required to understand how or know where that data may be stored. The service will deliver results using the latest up-to-date information – even on dynamic data stores – not just static reports or outdated indexes.
At the core of the solution is a large language model – Meta’s Llama2 or Cohere’s Command – that allows natural back-and-forth conversations with users. Think of it like a helpful colleague you can ask business questions in plain language. This is supplemented by the embeddings model that generates embeddings vectors for the corpus of corporate documents stored in the embeddings vector database. The retrieval-augmented generation (RAG) system is a combination of the vector database, embeddings model, and text generation model to synthesize grounded responses based on the documents retrieved. Instead of simply displaying the list of documents found, the AI service uses the text generation model to produce a natural language response that answers the original question, supported with references and links to original source documents. In the future, users will be able to not only retrieve information, but take immediate action such as sending emails, contacting customers, or managing orders directly from the same interface without needing to switch to other applications.
RAG provides the following benefits:
RAG offers different use cases for different industries and departments, such as the following examples:
The possibilities are vast for how this RAG service could transform knowledge acquisition and data-driven decision making. Whether it’s increasing sales opportunities, optimizing support resources, or simply empowering any employee with fast access to helpful information, the OCI Generative AI Agents RAG service can improve business productivity and customer experience.
For more information see the following resources:
Barry is a senior director for product marketing covering Oracle's AI and Analytics services.