Robuta

https://towardsdatascience.com/why-weve-been-optimizing-the-wrong-thing-in-llms-for-years/ Training Objective That Makes LLM Inference 3X Faster | Towards Data Science Dec 4, 2025 - The simple shift in training that unlocks foresight, faster inference, and better reasoning. llm inferencetowards data https://blog.exolabs.net/nvidia-dgx-spark/ Combining NVIDIA DGX Spark + Apple Mac Studio for 4x Faster LLM Inference with EXO 1.0 | EXO How to optimize both TTFT and TPS by splitting prefill and decode across different hardware nvidia dgx sparkapple mac https://predibase.com/blog/turbo-lora Turbo LoRA: 2-3x faster fine-tuned LLM inference Turbo LoRA is a new parameter-efficient fine-tuning method we’ve developed at Predibase that increases text generation throughput by 2-3x while... turbo lorafaster finetunedllm https://www.efficientlyconnected.com/rapidfire-ai-opens-new-era-of-llm-fine-tuning-with-open-source-engine/ RapidFire AI Open-Sources Engine for 20x Faster LLM Fine-Tuning Sep 29, 2025 - RapidFire AI releases an open-source ai openfaster llmrapidfire https://pulse2.com/rapidfire-ai-profile-arun-kumar-interview/ RapidFire AI: Interview With Co-Founder And CTO Arun Kumar About Making LLM Customization Faster RapidFire AI is a company that accelerates deep learning development by enabling data scientists to run many model and hyper-parameter experiments in parallel,... co founderarun kumarrapidfire https://predibase.com/blog/guide-how-to-serve-llms-faster-inference LLM Serving Guide: How to Build Faster Inference for Open-source Models Learn how to accelerate and optimize deployments for open-source models with our blueprint for fast, reliable, and cost-efficient LLM serving. Deep dive on GPU... build fasterllmservingguide