SynthesizeMe! Inducing Persona-Guided Prompts for Personalized Reward Models in LLMs Paper • 2506.05598 • Published Jun 5 • 7
Gemstone Models Collection Our 22 open source Gemstone models for scaling laws range from 50M to 2B parameters, spanning 11 widths from 256 to 3072 and 18 depths from 3 to 80. • 69 items • Updated Jul 4 • 10
Nemotron-CC: Transforming Common Crawl into a Refined Long-Horizon Pretraining Dataset Paper • 2412.02595 • Published Dec 3, 2024 • 5
Mind the Gap! Static and Interactive Evaluations of Large Audio Models Paper • 2502.15919 • Published Feb 21 • 4
view article Article Optimizing Pretraining Data Mixes with LLM-Estimated Utility By WillHeld • Jan 22 • 4
view article Article AI Apps in a Flash with Gradio's Reload Mode By freddyaboulton • Apr 16, 2024 • 27
Tutor CoPilot: A Human-AI Approach for Scaling Real-Time Expertise Paper • 2410.03017 • Published Oct 3, 2024 • 29
Distilling an End-to-End Voice Assistant Without Instruction Training Data Paper • 2410.02678 • Published Oct 3, 2024 • 23