Loading...
Corrective RAG (CRAG)(CRAG)
RAG system that automatically detects and corrects poor retrieval results through quality assessment and re-retrieval
๐ฏ 30-Second Overview
Pattern: Enhanced RAG with explicit retrieval quality evaluation and corrective actions based on confidence scoring
Why: Reduces hallucinations and improves accuracy through quality assessment and adaptive correction strategies
Key Insight: Three-tier correction strategy: refine high-confidence, supplement medium-confidence, re-retrieve low-confidence results
โก Quick Implementation
๐ Do's & Don'ts
๐ฆ When to Use
Use When
- โข High-stakes applications requiring verified accuracy and provenance
- โข Rapidly changing domains with frequent content updates
- โข Noisy or heterogeneous knowledge bases with quality variation
- โข Long-tail queries where initial retrieval often fails
- โข Regulated environments requiring explicit evidence grounding
Avoid When
- โข Real-time applications with strict latency requirements
- โข Closed-book tasks where parametric knowledge suffices
- โข High-quality homogeneous corpora with consistent recall
- โข Environments prohibiting external web access for supplementation
- โข Simple factual queries with reliable standard RAG performance
๐ Key Metrics
๐ก Top Use Cases
References & Further Reading
Deepen your understanding with these curated resources
Foundational Papers & CRAG Research
Corrective Retrieval Augmented Generation (Yan et al., 2024)
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks (Lewis et al., 2020)
Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection (Asai et al., 2023)
Active Retrieval Augmented Generation (Jiang et al., 2023)
Quality Assessment & Evaluation
Knowledge Refinement & Processing
Lost in the Middle: How Language Models Use Long Contexts (Liu et al., 2023)
LongLLMLingua: Accelerating Large Language Model Inference via Prompt Compression (Jiang et al., 2023)
Chain-of-Verification Reduces Hallucination in Large Language Models (Dhuliawala et al., 2023)
Factuality Enhanced Language Models for Open-Ended Text Generation (Lee et al., 2022)
Contribute to this collection
Know a great resource? Submit a pull request to add it.
Corrective RAG (CRAG)(CRAG)
RAG system that automatically detects and corrects poor retrieval results through quality assessment and re-retrieval
๐ฏ 30-Second Overview
Pattern: Enhanced RAG with explicit retrieval quality evaluation and corrective actions based on confidence scoring
Why: Reduces hallucinations and improves accuracy through quality assessment and adaptive correction strategies
Key Insight: Three-tier correction strategy: refine high-confidence, supplement medium-confidence, re-retrieve low-confidence results
โก Quick Implementation
๐ Do's & Don'ts
๐ฆ When to Use
Use When
- โข High-stakes applications requiring verified accuracy and provenance
- โข Rapidly changing domains with frequent content updates
- โข Noisy or heterogeneous knowledge bases with quality variation
- โข Long-tail queries where initial retrieval often fails
- โข Regulated environments requiring explicit evidence grounding
Avoid When
- โข Real-time applications with strict latency requirements
- โข Closed-book tasks where parametric knowledge suffices
- โข High-quality homogeneous corpora with consistent recall
- โข Environments prohibiting external web access for supplementation
- โข Simple factual queries with reliable standard RAG performance
๐ Key Metrics
๐ก Top Use Cases
References & Further Reading
Deepen your understanding with these curated resources
Foundational Papers & CRAG Research
Corrective Retrieval Augmented Generation (Yan et al., 2024)
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks (Lewis et al., 2020)
Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection (Asai et al., 2023)
Active Retrieval Augmented Generation (Jiang et al., 2023)
Quality Assessment & Evaluation
Knowledge Refinement & Processing
Lost in the Middle: How Language Models Use Long Contexts (Liu et al., 2023)
LongLLMLingua: Accelerating Large Language Model Inference via Prompt Compression (Jiang et al., 2023)
Chain-of-Verification Reduces Hallucination in Large Language Models (Dhuliawala et al., 2023)
Factuality Enhanced Language Models for Open-Ended Text Generation (Lee et al., 2022)
Contribute to this collection
Know a great resource? Submit a pull request to add it.