Running on CPU Upgrade Featured 3.18k The Smol Training Playbook 📚 3.18k The secrets to building world-class LLMs
Running 3.84k The Ultra-Scale Playbook 🌌 3.84k The ultimate guide to training LLM on large GPU Clusters
view article Article KV Caching Explained: Optimizing Transformer Inference Efficiency not-lain • Jan 30, 2025 • 329
Training Dynamics Impact Post-Training Quantization Robustness Paper • 2510.06213 • Published Oct 7, 2025 • 3
Training Dynamics Impact Post-Training Quantization Robustness Paper • 2510.06213 • Published Oct 7, 2025 • 3
view article Article Prefill and Decode for Concurrent Requests - Optimizing LLM Performance tngtech • Apr 16, 2025 • 78