Retrieval-Augmented Generation (RAG)

Definition

An AI architecture that combines a large language model with an external knowledge retrieval system, enabling the model to ground its responses in verified, up-to-date information rather than relying solely on its training data. RAG reduces hallucination risk, improves factual accuracy, and allows organisations to deploy AI systems that reference proprietary knowledge bases without retraining the underlying model.

Complementary Terms

Concepts that frequently appear alongside Retrieval-Augmented Generation (RAG) in practice.

Retrieval-Augmented Generation (RAG) Architecture

A technical architecture that enhances large language model outputs by retrieving relevant information from an external knowledge base before generating a response, grounding the model's output in verified, up-to-date, and domain-specific data. RAG reduces hallucination risk, enables LLMs to access proprietary or recent information not in their training data, and provides citation capabilities.

Model Drift

The degradation in a machine learning model's predictive accuracy over time as the statistical properties of the input data diverge from the training data distribution. Model drift requires ongoing monitoring and periodic retraining to maintain performance, and is a key operational risk in production AI systems.

AI Hallucination

An output generated by an artificial intelligence system — particularly large language models — that is factually incorrect, fabricated, or nonsensical, yet presented with apparent confidence. AI hallucinations pose significant risks in applications such as legal research, medical advice, and financial analysis, and their mitigation through grounding, retrieval-augmented generation, and human oversight is a key challenge in enterprise AI deployment.

Fine-Tuning

The process of further training a pre-trained machine learning model on a smaller, domain-specific dataset to adapt it for a particular task or industry. Fine-tuning allows organisations to leverage foundational models while creating proprietary, specialised AI capabilities that constitute identifiable intangible assets.

Data Lineage

The documented lifecycle of data as it moves through an organisation's systems, showing its origin, transformations, dependencies, and destinations. Data lineage provides visibility into how data is created, processed, and consumed, enabling organisations to ensure data quality, comply with regulatory requirements (particularly GDPR's right to explanation), debug data pipeline issues, and assess the impact of system changes.

Growth Forecasting

The process of projecting a company's future growth trajectory based on historical data, market conditions, and investment patterns. Incorporating intangible asset data and productivity trends significantly improves forecast accuracy and reduces investor uncertainty.

Firm-Specific Human Capital

The skills, knowledge, and expertise that are uniquely valuable within a specific organisation and less transferable to other employers. Firm-specific human capital is a critical intangible asset that grows through on-the-job training, institutional learning, and experience with proprietary systems and processes.

Large Language Model

A type of neural network trained on vast corpora of text data, capable of generating human-like text, answering questions, summarising documents, and performing reasoning tasks. Large language models such as GPT and Claude represent significant R&D investment and are reshaping knowledge work, customer service, and content production across industries.

Put this knowledge to work

Use Opagio's free tools to measure and grow the intangible assets that drive your business value.