Prompt Engineering Guide
😃 Basics
💼 Applications
🧙‍♂️ Intermediate
🧠 Advanced
Special Topics
🌱 New Techniques
🤖 Agents
⚖️ Reliability
🖼️ Image Prompting
🔓 Prompt Hacking
🔨 Tooling
💪 Prompt Tuning
🗂️ RAG
🎲 Miscellaneous
Models
📝 Language Models
Resources
📙 Vocabulary Resource
📚 Bibliography
📦 Prompted Products
🛸 Additional Resources
🔥 Hot Topics
✨ Credits

Introduction

🟢 This article is rated easy
Reading Time: 1 minute

Last updated on August 7, 2024

This chapter covers Retrieval-Augmented Generation (RAG) techniques that enhances the capabilities of Large Language Models (LLMs) by integrating external data retrieval.

Instead of relying solely on pre-trained data, RAG allows models to fetch relevant information from external sources, making them more dynamic and adaptable. This architecture solves key limitations of LLMs, such as "hallucinations" (generating false information) and the need for constant fine-tuning to reflect updated data.

Valeriia Kuka

Valeriia Kuka, Head of Content at Learn Prompting, is passionate about making AI and ML accessible. Valeriia previously grew a 60K+ follower AI-focused social media account, earning reposts from Stanford NLP, Amazon Research, Hugging Face, and AI researchers. She has also worked with AI/ML newsletters and global communities with 100K+ members and authored clear and concise explainers and historical articles.

🟦 Multi-Fusion Retrieval Augmented Generation (MoRAG)

Edit this page
Word count: 0

Get AI Certified by Learn Prompting


Copyright © 2024 Learn Prompting.