Robuta

Sponsor of the Day: Jerkmate
https://snowflake.discourse.group/u/Run_LLM_Agent Profile - Run_LLM_Agent - Snowflake run llmprofileagentsnowflake https://www.hongkiat.com/blog/ollama-llm-from-external-drive/ How to Run LLM from External Hard Drive with Ollama - Hongkiat Want to run generative AI right on your Mac? There's where Ollama comes in. But there's a problem - modern LLMs can eat up anywhere from 4GB to 40GB each, external hard driverun llmollama hongkiat https://www.hongkiat.com/blog/docker-llm-setup-guide/ How to Run LLM in Docker - Hongkiat Large Language Models (LLMs) have changed how we build and use software. While cloud-based LLM APIs are great for convenience, there are plenty of reasons run llmdockerhongkiat https://simonwillison.net/2026/Mar/18/llm-in-a-flash/ Autoresearching Apple's "LLM in a Flash" to run Qwen 397B locally Here's a fascinating piece of research by Dan Woods, who managed to get a custom version of Qwen3.5-397B-A17B running at 5.5+ tokens/second on a 48GB MacBook... applellmflashrunqwen https://simonwillison.net/2025/Feb/15/llm-mlx/ Run LLMs on macOS using llm-mlx and Apple’s MLX framework llm-mlx is a brand new plugin for my LLM Python Library and CLI utility which builds on top of Apple’s excellent MLX array framework library and mlx-lm... run llmsmacosusingmlxframework https://www.bcdvideo.com/blog/what-hardware-is-needed-to-run-llms/ What Hardware Do You Need to Run an LLM? | BCD Apr 7, 2026 - Interested in running a local large language model? Here's everything you should know about the various hardware you'll need for a high-performing LLM. hardwareneedrunllmbcd https://apxml.com/tools/vram-calculator Can You Run This LLM? VRAM Calculator (Nvidia GPU and Apple Silicon) Calculate the VRAM required to run any large language model. nvidia gpuapple siliconrunllmvram https://simonwillison.net/2024/Oct/27/llm-jq/ Run a prompt to generate and execute jq programs using llm-jq llm-jq is a brand new plugin for LLM which lets you pipe JSON directly into the llm jq command along with a human-language description of how you’d like to... programs usingrunpromptgenerateexecute https://www.xda-developers.com/gemma-4-not-smartest-local-llm-but-reach-for/ Google's Gemma 4 isn't the smartest local LLM I've run, but it's the one I reach for most Apr 15, 2026 - Google's newest Gemma 4 models are both powerful and useful. gemma 4local llmgooglesmartestrun https://novoserve.com/blog/how-to-run-an-llm-on-a-server-your-2026-llm-server-hardware-guide How to Run an LLM on a Server: Your 2026 LLM Server Hardware Guide Jan 21, 2026 - The rise of Large Language Models (LLMs) has been transformative, but how do you run an LLM on a server on your own? 2026 hardwarerunllmserverguide https://simonwillison.net/2025/May/27/llm-tools/ Large Language Models can run tools in your terminal with LLM 0.26 LLM 0.26 is out with the biggest new feature since I started the project: support for tools. You can now use the LLM CLI tool—and Python library—to grant LLMs... large language modelsllm 0runtoolsterminal https://www.makeuseof.com/you-can-and-should-run-a-tiny-llm-on-your-android-phone/ You can (and should) run a tiny LLM on your Android phone Feb 9, 2026 - Unused RAM is wasted RAM android phoneruntinyllm https://thenewstack.io/how-to-set-up-and-run-a-local-llm-with-ollama-and-llama-2/ How To Use Ollama: Set Up and Run a Local LLM With Llama 3 Jan 31, 2025 - Take a look at how to run an open source LLM locally, which allows you to run queries on your private data without any security concerns. use ollamalocal llmsetrun3