Martin Keen, a Master Inventor at IBM, breaks down the fundamental challenges and solutions surrounding the integration of external data into Large Language Models (LLMs). Keen highlights the inherent limitation of LLMs: their knowledge is frozen at the time of their last training data cutoff. This means they cannot access or process information that has emerged since then, creating a significant hurdle for applications requiring up-to-date or specific context.
Understanding LLM Limitations
Keen explains that LLMs, by their nature, are static models. They possess all the knowledge they were trained on, but they have no awareness of anything that has happened since their training data was collected. This static nature presents a problem when users need LLMs to interact with current events, proprietary company data, or any information not included in the original training corpus. To address this, two primary approaches have emerged: Retrieval Augmented Generation (RAG) and simply increasing the context window size.
Retrieval Augmented Generation (RAG) Explained
Keen illustrates the RAG process, which he describes as a foundational truth about how we get the right data into an LLM at the right time. The RAG approach involves several key steps:
