- Turing Post
- Posts
- Topic 27: What are Chain-of-Agents and Chain-of-RAG?
Topic 27: What are Chain-of-Agents and Chain-of-RAG?
We explore Google's and Microsoft's advancements that implement "chain" approaches for long context and multi-hop reasoning
With the shift to deep, step-by-step reasoning in AI, we continue to observe a trend of creating Chain-of-ââŠâ methods. Previously, we explored three Chains-of-Knowledge and other "chain" spin-offs in âFrom Chain-of-Thoughts to Skeleton-of-Thoughts, and everything in betweenâ, but âchainsâ keep coming! Today, weâre going to discuss the two advancements from Google Cloud AI Research and Microsoft, called Chain-of-Agents (CoA) and Chain-of-Retrieval Augmented Generation (CoRAG), respectively. They both approach the challenge of handling long-context tasks, but from different perspectives. Googleâs CoA employs a multi-agent framework, where worker agents process text segments sequentially in a structured chain, while Microsoftâs CoRAG introduces an iterative retrieval approach as a solution for strong multi-hop reasoning. Understanding techniques like CoA and CoRAG is crucial if you are working toward improving AI's performance in complex reasoning tasks. So, letâs explore how these new âchainsâ can impact the accuracy and quality of AI models!
In todayâs episode, we will cover:
Chain-of-Agents from Google: whatâs the idea?
How does CoA work?
How good CoA actually is?
CoA advantages and why it is better than RAG and other methods
Not without limitations
The key idea of Chain-of-RAG (CoRAG) from Microsoft
How does CoRAG work?
Performance of CoRAG
CoRAGâs advantages
Not without limitations
Conclusion
Bonus: Resources to dive deeper
Chain-of-Agents from Google: whatâs the idea?
Even when you are working with state-of-the-art models, you can notice that tasks with long context, like entire books, long articles, or lengthy conversations, still remain a challenge for LLMs. One of the widespread ideas is to expand the modelâs memory, in other words, context window. However, models tend to lose the track of main information as the input grows longer. Another way is to shorten input instead by selecting only the most relevant parts of the text. Here RAG may be used for effective retrieval, but this method may lead to losing important parts of information.
What to do? Google Cloud AI Research and Penn State Universityâs researchers pursued another strategy to create a method that would be better than RAG, full-context models, and multi-agent LLMs. They proposed the Chain-of-Agents (CoA) framework, inspired by how humans process long texts step-by-step. Instead of relying on a single model, CoA enables multiple AI agents to collaborate and process unlimited amounts of text.
Collaboration among agents may not seem a new concept, but researchers have discovered some tips that make their method stand out. Many methods use tree structure where agents work separately without direct communication (for example, LongAgent). In contrast, CoA follows a chain structure with a strong order, at the same time ensuring agents share information for better accuracy. Letâs look at how CoA exactly does it.
How does CoA work?

Reply