Running on CPU Upgrade Featured 2.72k The Smol Training Playbook 📚 2.72k The secrets to building world-class LLMs
Running 3.6k The Ultra-Scale Playbook 🌌 3.6k The ultimate guide to training LLM on large GPU Clusters
view article Article KV Caching Explained: Optimizing Transformer Inference Efficiency Jan 30 • 202
Training Dynamics Impact Post-Training Quantization Robustness Paper • 2510.06213 • Published Oct 7 • 3
Training Dynamics Impact Post-Training Quantization Robustness Paper • 2510.06213 • Published Oct 7 • 3
view article Article Prefill and Decode for Concurrent Requests - Optimizing LLM Performance Apr 16 • 56
🧠SmolLM3 Collection Smol, multilingual, long-context reasoner • 14 items • Updated Oct 9 • 89
view article Article Unlocking Longer Generation with Key-Value Cache Quantization May 16, 2024 • 54