OLMoTrace: Tracing Language Model Outputs Back to Trillions of Training Tokens Paper • 2504.07096 • Published Apr 9 • 76
Hybrid Preferences: Learning to Route Instances for Human vs. AI Feedback Paper • 2410.19133 • Published Oct 24, 2024 • 11
Null It Out: Guarding Protected Attributes by Iterative Nullspace Projection Paper • 2004.07667 • Published Apr 16, 2020
Few-shot Fine-tuning vs. In-context Learning: A Fair Comparison and Evaluation Paper • 2305.16938 • Published May 26, 2023
Lexical Generalization Improves with Larger Models and Longer Training Paper • 2210.12673 • Published Oct 23, 2022
Data Contamination Report from the 2024 CONDA Shared Task Paper • 2407.21530 • Published Jul 31, 2024 • 10
Dolma: an Open Corpus of Three Trillion Tokens for Language Model Pretraining Research Paper • 2402.00159 • Published Jan 31, 2024 • 65
Paloma: A Benchmark for Evaluating Language Model Fit Paper • 2312.10523 • Published Dec 16, 2023 • 13