- What Language Model Architecture and Pretraining Objective Work Best for Zero-Shot Generalization
- HuggingFace Tasks
- Model Hub
- LLaMA: Open and Efficient Foundation Language Models
- Language Models are Few-Shot Learners
- Training Compute-Optimal Large Language Models
- BloombergGPT: A Large Language Model for Finance
- Scaling Instruction-Finetuned Language Models
- Introducing FLAN: More generalizable Language Models with Instruction Fine-Tuning
- HELM - Holistic Evaluation of Language Models
- General Language Understanding Evaluation (GLUE) benchmark
- SuperGLUE
- ROUGE: A Package for Automatic Evaluation of Summaries
- Measuring Massive Multitask Language Understanding (MMLU)
- BigBench-Hard - Beyond the Imitation Game: Quantifying and Extrapolating the Capabilities of Language Models
- Scaling Down to Scale Up: A Guide to Parameter-Efficient Fine-Tuning
- On the Effectiveness of Parameter-Efficient Fine-Tuning
- Training language models to follow instructions with human feedback
- Learning to summarize from human feedback
- Proximal Policy Optimization Algorithms
- Direct Preference Optimization: Your Language Model is Secretly a Reward Model
- Chain-of-thought Prompting Elicits Reasoning in Large Language Models
- PAL: Program-aided Language Models
- ReAct: Synergizing Reasoning and Acting in Language Models
Credits: