RAG is Retrieval Augmented Generation. It is a fancy way of saying “we’ve tacked a search engine onto the LLM so that it can query for and use the text of actual documents when generating text, so that the output is more likely to be correct and grounded in reality.”
And yeah, MCP stands for Model Context Protocol, and is essentially an API format optimized for LLMs, as you’ve said, to defer to something else to do the work. This can be a (RAG like) search engine lookup, using a calculator, or something else entirely.
LLMs suck at doing a lot of stuff reliably (like calculations, making statements relating to recent events, …), but they turn out to be quite a useful tool for translating between human and machine, and reasonably capable of stringing things together to get an answer.
RAG is Retrieval Augmented Generation. It is a fancy way of saying “we’ve tacked a search engine onto the LLM so that it can query for and use the text of actual documents when generating text, so that the output is more likely to be correct and grounded in reality.”
And yeah, MCP stands for Model Context Protocol, and is essentially an API format optimized for LLMs, as you’ve said, to defer to something else to do the work. This can be a (RAG like) search engine lookup, using a calculator, or something else entirely.
LLMs suck at doing a lot of stuff reliably (like calculations, making statements relating to recent events, …), but they turn out to be quite a useful tool for translating between human and machine, and reasonably capable of stringing things together to get an answer.