Ep67: Why RAG Fails LLMs – And How to Finally Fix It

19/03/2025 22 min Temporada 2 Episodio 67

Listen "Ep67: Why RAG Fails LLMs – And How to Finally Fix It"

Episode Synopsis

AI is lying to you—here’s why. Retrieval-Augmented Generation (RAG) was supposed to fix AI hallucinations, but it’s failing. In this episode, we break down the limitations of naïve RAG, the rise of dense retrieval, and how new approaches like Agentic RAG, RePlug, and RAG Fusion are revolutionizing AI search accuracy.🔍 Key Insights:Why naïve RAG fails and leads to bad retrievalHow Contriever & Dense Retrieval improve accuracyRePlug’s approach to refining AI queriesWhy RAG Fusion is a game-changer for AI searchThe future of AI retrieval beyond vector databasesIf you’ve ever wondered why LLMs still struggle with real knowledge retrieval, this is the episode you need!🎧 Listen now and stay ahead in AI!References:[2005.11401] Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks[2112.09118] Unsupervised Dense Information Retrieval with Contrastive Learning[2301.12652] REPLUG: Retrieval-Augmented Black-Box Language Models[2402.03367] RAG-Fusion: a New Take on Retrieval-Augmented Generation[2312.10997] Retrieval-Augmented Generation for Large Language Models: A Survey

More episodes of the podcast Machine Learning Made Simple