📢 Top RAG Papers of the Week (December Week 1, 2024)
[1] Impact of OCR on RAG
This paper introduces OHRBench for understanding the impact of OCR on RAG systems. OHRBench includes 350 carefully selected unstructured PDF documents from six real-world RAG application domains. The authors systematically evaluate the impact of these two noise types (Semantic Noise and Formatting Noise) and demonstrate the vulnerability of RAG systems. [Tweet] and [Paper]
[2] Auto-RAG
This paper introduces Auto-RAG, autonomous RAG for LLMs. Auto-RAG engages in multi-turn dialogues with the retriever, systematically planning retrievals and refining queries to acquire valuable knowledge. Auto-RAG achieves outstanding performances across six benchmarks because of the ability to autonomously interact with the retriever and effectively leverage the decision-making abilities of LLMs. [Tweet] and [Paper]
[3]Analyzing OpenAPI Chunking for RAG
This paper investigates OpenAPI chunking for RAG and addresses the question, “Can LLM agents be employed to reduce token count further and improve retrieval performance?”. For this, the authors a Discovery Agent that only receives a summary of the most relevant endpoints and retrieves details on demand. Results show that (i) LLM-based and format-specific chunking outperform naïve chunking methods. (ii) Relying on an agent further enhances these results as the agent splits the tasks into multiple fine granular subtasks. [Tweet] and [Paper]
Recommended by LinkedIn
[4] Know Your RAG
This paper shows that (i) using public question and answer (Q&A) datasets to assess retrieval performance can lead to non-optimal systems design, and (ii) common tools for RAG dataset generation can lead to unbalanced data. The authors (i) propose solutions to these issues based on the characterization of RAG datasets through labels and through label-targeted data generation and (ii) show that fine-tuned small LLMs can efficiently generate Q&A datasets. [Tweet] and [Paper]
[5] Understanding Retrieval Accuracy and Prompt Quality in RAG Systems
This paper presents a study to understand retrieval accuracy and prompt quality in RAG systems by conducting experiments on three code datasets, three QA datasets, and two LLMs. The authors focus on four design factors: retrieval document type, retrieval recall, document selection, and prompt techniques. Based on the results, they present nine actionable guidelines for detecting defects and optimizing the performance of RAG systems. [Tweet] and [Paper]
[6] MBA-RAG
The paper introduces MBA-RAG which uses a reinforcement learning-based framework that dynamically selects the most suitable retrieval strategy based on query complexity. The approach leverages a multi-armed bandit algorithm, which treats each retrieval method as a distinct “arm” and adapts the selection process by balancing exploration and exploitation. MBA-RAG achieves new state-of-the-art results on multiple single-hop and multi-hop datasets while reducing retrieval costs. [Tweet] and [Paper]
Do subscribe to the newsletter so that you won't miss interesting updates related to Generative AI, LLMs, Agents and RAG.
Kalyan KS, Research Scientist(NLP) at Akmmus AI Labs
Data Analyst at WorkL | 4+ Years in Data Engineering | Master's in Data Science | Skilled in Python, SQL, and Machine Learning | Enthusiast in NLP
2wThanks for sharing:)
Senior AI Engineer | AI/ML/LLM Developer | IIT Hyderabad | JNV | Navodaya
2wThanks for sharing
Senior Project Manager at Infosys
2wThanks for sharing
Head of Technology @21 Spheres || IIT(BHU) Varanasi
2wVery informative
Vector Compute @ Superlinked | xYouTube
2wnice initiative