Robuta

https://northflank.com/stacks/deploy-litellm
LiteLLM is a lightweight, self-hosted LLM server for secure, real-time inference. Deploy it on Northflank in minutes for a scalable, production-ready...
deploylitellmstacktemplates
https://langfuse.com/integrations/gateways/litellm
open sourceobservabilitylitellmproxylangfuse
https://developers.freee.co.jp/entry/aws-builders-flash-202511
Nov 5, 2025 - 弊社のAI駆動開発チームの中山が、Amazon Web Service (AWS) のブログメディアである、Builders Flash にて記事を寄稿しました。...
buildersflashawslitellm
https://www.linux.com/news/how-to-deploy-lightweight-language-models-on-embedded-linux-with-litellm/
Jun 6, 2025 - This article was contributed by Vedrana Vidulin, Head of Responsible AI Unit at Intellias (LinkedIn). As AI becomes central to smart devices, embedded systems,...
language modelsembedded linuxdeploylightweight
https://distilabel.argilla.io/latest/components-gallery/llms/litellm/
Distilabel is an AI Feedback (AIF) framework for building datasets with and for LLMs.
litellmdocs
https://dev.to/debmckinney/litellm-broke-at-300-rps-in-production-heres-how-we-fixed-it-5ej
Jan 9, 2026 - The Incident Tuesday, 2:47 PM Our customer support chatbot is handling 280 RPS.... Tagged with go, llm, chatgpt, ai.
litellmbrokerpsproduction