-
Ensemble-Instruct: Generating Instruction-Tuning Data with a Heterogeneous Mixture of LMs
Paper β’ 2310.13961 β’ Published β’ 4 -
ZeroGen: Efficient Zero-shot Learning via Dataset Generation
Paper β’ 2202.07922 β’ Published β’ 1 -
Let's Synthesize Step by Step: Iterative Dataset Synthesis with Large Language Models by Extrapolating Errors from Small Models
Paper β’ 2310.13671 β’ Published β’ 18 -
Fabricator: An Open Source Toolkit for Generating Labeled Training Data with Teacher LLMs
Paper β’ 2309.09582 β’ Published β’ 4
Collections
Discover the best community collections!
Collections including paper arxiv:2306.11644
-
A Survey on Language Models for Code
Paper β’ 2311.07989 β’ Published β’ 21 -
Evaluating Large Language Models Trained on Code
Paper β’ 2107.03374 β’ Published β’ 8 -
SWE-bench: Can Language Models Resolve Real-World GitHub Issues?
Paper β’ 2310.06770 β’ Published β’ 4 -
CodeXGLUE: A Machine Learning Benchmark Dataset for Code Understanding and Generation
Paper β’ 2102.04664 β’ Published β’ 2
-
AgentInstruct: Toward Generative Teaching with Agentic Flows
Paper β’ 2407.03502 β’ Published β’ 51 -
Magpie: Alignment Data Synthesis from Scratch by Prompting Aligned LLMs with Nothing
Paper β’ 2406.08464 β’ Published β’ 65 -
Phi-3 Technical Report: A Highly Capable Language Model Locally on Your Phone
Paper β’ 2404.14219 β’ Published β’ 253 -
DataDreamer: A Tool for Synthetic Data Generation and Reproducible LLM Workflows
Paper β’ 2402.10379 β’ Published β’ 30
-
SciLitLLM: How to Adapt LLMs for Scientific Literature Understanding
Paper β’ 2408.15545 β’ Published β’ 34 -
Controllable Text Generation for Large Language Models: A Survey
Paper β’ 2408.12599 β’ Published β’ 63 -
To Code, or Not To Code? Exploring Impact of Code in Pre-training
Paper β’ 2408.10914 β’ Published β’ 41 -
Automated Design of Agentic Systems
Paper β’ 2408.08435 β’ Published β’ 38
-
Textbooks Are All You Need
Paper β’ 2306.11644 β’ Published β’ 142 -
Textbooks Are All You Need II: phi-1.5 technical report
Paper β’ 2309.05463 β’ Published β’ 87 -
TinyStories: How Small Can Language Models Be and Still Speak Coherent English?
Paper β’ 2305.07759 β’ Published β’ 33 -
Scaling Synthetic Data Creation with 1,000,000,000 Personas
Paper β’ 2406.20094 β’ Published β’ 96
-
Rephrasing the Web: A Recipe for Compute and Data-Efficient Language Modeling
Paper β’ 2401.16380 β’ Published β’ 48 -
Best Practices and Lessons Learned on Synthetic Data for Language Models
Paper β’ 2404.07503 β’ Published β’ 29 -
WizardLM: Empowering Large Language Models to Follow Complex Instructions
Paper β’ 2304.12244 β’ Published β’ 13 -
Synthetic Data (Almost) from Scratch: Generalized Instruction Tuning for Language Models
Paper β’ 2402.13064 β’ Published β’ 47