Loading...
Self-RAG(SRAG)
Self-reflective RAG that adaptively determines retrieval necessity and evaluates retrieval quality through reflection tokens
๐ฏ 30-Second Overview
Pattern: RAG system with adaptive retrieval decisions and self-reflection using trained reflection tokens for quality assessment
Why: Improves factual accuracy and reduces hallucinations through iterative self-critique and selective knowledge retrieval
Key Insight: Reflection tokens ([Retrieve], [IsRel], [IsSup], [IsUse]) enable models to assess retrieval necessity and response quality
โก Quick Implementation
๐ Do's & Don'ts
๐ฆ When to Use
Use When
- โข Factual accuracy and verifiability are critical requirements
- โข Domain expertise requires balancing parametric and retrieved knowledge
- โข Applications need confidence calibration and uncertainty quantification
- โข High-stakes decisions requiring explainable reasoning and citations
- โข Knowledge-intensive tasks in medical, legal, or scientific domains
Avoid When
- โข Simple queries where standard RAG provides sufficient accuracy
- โข Real-time applications with strict latency constraints
- โข Resource-constrained environments limiting multiple generation rounds
- โข Domains with insufficient training data for reliable self-critique
- โข Applications where citation overhead is unnecessary
๐ Key Metrics
๐ก Top Use Cases
References & Further Reading
Deepen your understanding with these curated resources
Foundational Papers & Self-RAG Research
Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection (Asai et al., 2023)
SELF-REFINE: Iterative Refinement with Self-Feedback (Madaan et al., 2023)
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks (Lewis et al., 2020)
Teaching Language Models to Self-Correct via Reinforcement Learning (Welleck et al., 2023)
Reflection Tokens & Training Methods
Retrieval & Context Assessment
When Not to Trust Language Models: Investigating Effectiveness of Parametric and Non-Parametric Memories (Mallen et al., 2023)
Active Retrieval Augmented Generation (Jiang et al., 2023)
FiD: Fusion-in-Decoder for Open-Domain Question Answering (Izacard & Grave, 2021)
Dense Passage Retrieval for Open-Domain Question Answering (Karpukhin et al., 2020)
Evaluation & Calibration Methods
Contribute to this collection
Know a great resource? Submit a pull request to add it.
Self-RAG(SRAG)
Self-reflective RAG that adaptively determines retrieval necessity and evaluates retrieval quality through reflection tokens
๐ฏ 30-Second Overview
Pattern: RAG system with adaptive retrieval decisions and self-reflection using trained reflection tokens for quality assessment
Why: Improves factual accuracy and reduces hallucinations through iterative self-critique and selective knowledge retrieval
Key Insight: Reflection tokens ([Retrieve], [IsRel], [IsSup], [IsUse]) enable models to assess retrieval necessity and response quality
โก Quick Implementation
๐ Do's & Don'ts
๐ฆ When to Use
Use When
- โข Factual accuracy and verifiability are critical requirements
- โข Domain expertise requires balancing parametric and retrieved knowledge
- โข Applications need confidence calibration and uncertainty quantification
- โข High-stakes decisions requiring explainable reasoning and citations
- โข Knowledge-intensive tasks in medical, legal, or scientific domains
Avoid When
- โข Simple queries where standard RAG provides sufficient accuracy
- โข Real-time applications with strict latency constraints
- โข Resource-constrained environments limiting multiple generation rounds
- โข Domains with insufficient training data for reliable self-critique
- โข Applications where citation overhead is unnecessary
๐ Key Metrics
๐ก Top Use Cases
References & Further Reading
Deepen your understanding with these curated resources
Foundational Papers & Self-RAG Research
Self-RAG: Learning to Retrieve, Generate, and Critique through Self-Reflection (Asai et al., 2023)
SELF-REFINE: Iterative Refinement with Self-Feedback (Madaan et al., 2023)
Retrieval-Augmented Generation for Knowledge-Intensive NLP Tasks (Lewis et al., 2020)
Teaching Language Models to Self-Correct via Reinforcement Learning (Welleck et al., 2023)
Reflection Tokens & Training Methods
Retrieval & Context Assessment
When Not to Trust Language Models: Investigating Effectiveness of Parametric and Non-Parametric Memories (Mallen et al., 2023)
Active Retrieval Augmented Generation (Jiang et al., 2023)
FiD: Fusion-in-Decoder for Open-Domain Question Answering (Izacard & Grave, 2021)
Dense Passage Retrieval for Open-Domain Question Answering (Karpukhin et al., 2020)
Evaluation & Calibration Methods
Contribute to this collection
Know a great resource? Submit a pull request to add it.