Loading...
Supervised Learning for Agents(SLA)
Learning from labeled examples to improve agent decision-making and task performance
๐ฏ 30-Second Overview
Pattern: Adapt pre-trained models to specific tasks through supervised training on labeled target domain data
Why: Leverages existing model knowledge while specializing for target tasks, achieving high performance with reasonable training costs
Key Insight: Transfer learning from pre-trained models combined with task-specific supervision provides optimal balance of generalization and specialization
โก Quick Implementation
๐ Do's & Don'ts
๐ฆ When to Use
Use When
- โข High-quality labeled data is available for target task
- โข Task has clear input-output relationships
- โข Performance requirements justify supervised training costs
- โข Domain-specific adaptation is needed from general models
- โข Evaluation metrics can be clearly defined
Avoid When
- โข Labeled data is scarce, expensive, or low quality
- โข Task requires real-time learning from minimal examples
- โข Unsupervised or self-supervised approaches are sufficient
- โข Privacy constraints prevent data collection
- โข Deployment environment changes frequently
๐ Key Metrics
๐ก Top Use Cases
References & Further Reading
Deepen your understanding with these curated resources
Foundational Papers
Fine-Tuning Pre-trained Language Models: Weight Initializations, Data Orders, and Early Stopping (Dodge et al., 2020)
How to Fine-Tune BERT for Text Classification? (Sun et al., 2019)
Universal Language Model Fine-tuning for Text Classification (Howard & Ruder, 2018)
Attention Is All You Need (Vaswani et al., 2017)
Transfer Learning & Pre-training
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (Devlin et al., 2018)
Language Models are Few-Shot Learners (Brown et al., 2020)
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (Raffel et al., 2019)
RoBERTa: A Robustly Optimized BERT Pretraining Approach (Liu et al., 2019)
Fine-Tuning Methodologies
LoRA: Low-Rank Adaptation of Large Language Models (Hu et al., 2021)
Prefix-Tuning: Optimizing Continuous Prompts for Generation (Li & Liang, 2021)
The Power of Scale for Parameter-Efficient Prompt Tuning (Lester et al., 2021)
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (Zhang et al., 2023)
Domain Adaptation Techniques
Domain-Adversarial Training of Neural Networks (Ganin et al., 2016)
Deep Domain Confusion: Maximizing for Domain Invariance (Tzeng et al., 2014)
Unsupervised Domain Adaptation by Backpropagation (Ganin & Lempitsky, 2015)
AdaBound: Adaptive Gradient Methods with Bound for Domain Adaptation (Luo et al., 2019)
Recent Advances (2023-2024)
QLoRA: Efficient Finetuning of Quantized LLMs (Dettmers et al., 2023)
LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention (Zhang et al., 2023)
Instruction Tuning for Large Language Models: A Survey (Zhang et al., 2023)
DoRA: Weight-Decomposed Low-Rank Adaptation (Liu et al., 2024)
Data Efficiency & Few-Shot Learning
Evaluation & Benchmarking
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding (Wang et al., 2018)
SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems (Wang et al., 2019)
BIG-bench: Beyond the Imitation Game Benchmark (Srivastava et al., 2022)
HELM: Holistic Evaluation of Language Models (Liang et al., 2022)
Contribute to this collection
Know a great resource? Submit a pull request to add it.
Supervised Learning for Agents(SLA)
Learning from labeled examples to improve agent decision-making and task performance
๐ฏ 30-Second Overview
Pattern: Adapt pre-trained models to specific tasks through supervised training on labeled target domain data
Why: Leverages existing model knowledge while specializing for target tasks, achieving high performance with reasonable training costs
Key Insight: Transfer learning from pre-trained models combined with task-specific supervision provides optimal balance of generalization and specialization
โก Quick Implementation
๐ Do's & Don'ts
๐ฆ When to Use
Use When
- โข High-quality labeled data is available for target task
- โข Task has clear input-output relationships
- โข Performance requirements justify supervised training costs
- โข Domain-specific adaptation is needed from general models
- โข Evaluation metrics can be clearly defined
Avoid When
- โข Labeled data is scarce, expensive, or low quality
- โข Task requires real-time learning from minimal examples
- โข Unsupervised or self-supervised approaches are sufficient
- โข Privacy constraints prevent data collection
- โข Deployment environment changes frequently
๐ Key Metrics
๐ก Top Use Cases
References & Further Reading
Deepen your understanding with these curated resources
Foundational Papers
Fine-Tuning Pre-trained Language Models: Weight Initializations, Data Orders, and Early Stopping (Dodge et al., 2020)
How to Fine-Tune BERT for Text Classification? (Sun et al., 2019)
Universal Language Model Fine-tuning for Text Classification (Howard & Ruder, 2018)
Attention Is All You Need (Vaswani et al., 2017)
Transfer Learning & Pre-training
BERT: Pre-training of Deep Bidirectional Transformers for Language Understanding (Devlin et al., 2018)
Language Models are Few-Shot Learners (Brown et al., 2020)
T5: Exploring the Limits of Transfer Learning with a Unified Text-to-Text Transformer (Raffel et al., 2019)
RoBERTa: A Robustly Optimized BERT Pretraining Approach (Liu et al., 2019)
Fine-Tuning Methodologies
LoRA: Low-Rank Adaptation of Large Language Models (Hu et al., 2021)
Prefix-Tuning: Optimizing Continuous Prompts for Generation (Li & Liang, 2021)
The Power of Scale for Parameter-Efficient Prompt Tuning (Lester et al., 2021)
AdaLoRA: Adaptive Budget Allocation for Parameter-Efficient Fine-Tuning (Zhang et al., 2023)
Domain Adaptation Techniques
Domain-Adversarial Training of Neural Networks (Ganin et al., 2016)
Deep Domain Confusion: Maximizing for Domain Invariance (Tzeng et al., 2014)
Unsupervised Domain Adaptation by Backpropagation (Ganin & Lempitsky, 2015)
AdaBound: Adaptive Gradient Methods with Bound for Domain Adaptation (Luo et al., 2019)
Recent Advances (2023-2024)
QLoRA: Efficient Finetuning of Quantized LLMs (Dettmers et al., 2023)
LLaMA-Adapter: Efficient Fine-tuning of Language Models with Zero-init Attention (Zhang et al., 2023)
Instruction Tuning for Large Language Models: A Survey (Zhang et al., 2023)
DoRA: Weight-Decomposed Low-Rank Adaptation (Liu et al., 2024)
Data Efficiency & Few-Shot Learning
Evaluation & Benchmarking
GLUE: A Multi-Task Benchmark and Analysis Platform for Natural Language Understanding (Wang et al., 2018)
SuperGLUE: A Stickier Benchmark for General-Purpose Language Understanding Systems (Wang et al., 2019)
BIG-bench: Beyond the Imitation Game Benchmark (Srivastava et al., 2022)
HELM: Holistic Evaluation of Language Models (Liang et al., 2022)
Contribute to this collection
Know a great resource? Submit a pull request to add it.