https://towardsdatascience.com/why-weve-been-optimizing-the-wrong-thing-in-llms-for-years/
Training Objective That Makes LLM Inference 3X Faster | Towards Data Science
Dec 4, 2025 - The simple shift in training that unlocks foresight, faster inference, and better reasoning.
llm inferencetowards data
https://blog.exolabs.net/nvidia-dgx-spark/
Combining NVIDIA DGX Spark + Apple Mac Studio for 4x Faster LLM Inference with EXO 1.0 | EXO
How to optimize both TTFT and TPS by splitting prefill and decode across different hardware
nvidia dgx sparkapple mac
https://predibase.com/blog/turbo-lora
Turbo LoRA: 2-3x faster fine-tuned LLM inference
Turbo LoRA is a new parameter-efficient fine-tuning method we’ve developed at Predibase that increases text generation throughput by 2-3x while...
turbo lorafaster finetunedllm
https://www.efficientlyconnected.com/rapidfire-ai-opens-new-era-of-llm-fine-tuning-with-open-source-engine/
RapidFire AI Open-Sources Engine for 20x Faster LLM Fine-Tuning
Sep 29, 2025 - RapidFire AI releases an open-source
ai openfaster llmrapidfire
https://pulse2.com/rapidfire-ai-profile-arun-kumar-interview/
RapidFire AI: Interview With Co-Founder And CTO Arun Kumar About Making LLM Customization Faster
RapidFire AI is a company that accelerates deep learning development by enabling data scientists to run many model and hyper-parameter experiments in parallel,...
co founderarun kumarrapidfire
https://predibase.com/blog/guide-how-to-serve-llms-faster-inference
LLM Serving Guide: How to Build Faster Inference for Open-source Models
Learn how to accelerate and optimize deployments for open-source models with our blueprint for fast, reliable, and cost-efficient LLM serving. Deep dive on GPU...
build fasterllmservingguide