-
InfiniteHiP: Extending Language Model Context Up to 3 Million Tokens on a Single GPU
Paper • 2502.08910 • Published • 149 -
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling
Paper • 2502.06703 • Published • 154 -
The Curse of Depth in Large Language Models
Paper • 2502.05795 • Published • 41
Shenxin Li
Adinosaur
·
AI & ML interests
None yet
Organizations
LLM structure optimization
-
InfiniteHiP: Extending Language Model Context Up to 3 Million Tokens on a Single GPU
Paper • 2502.08910 • Published • 149 -
Can 1B LLM Surpass 405B LLM? Rethinking Compute-Optimal Test-Time Scaling
Paper • 2502.06703 • Published • 154 -
The Curse of Depth in Large Language Models
Paper • 2502.05795 • Published • 41
LLM Evaluation