Sponsor of the Day:
Jerkmate
https://snowflake.discourse.group/u/Run_LLM_Agent
Profile - Run_LLM_Agent - Snowflake
run llmprofileagentsnowflake
https://www.hongkiat.com/blog/ollama-llm-from-external-drive/
How to Run LLM from External Hard Drive with Ollama - Hongkiat
Want to run generative AI right on your Mac? There's where Ollama comes in. But there's a problem - modern LLMs can eat up anywhere from 4GB to 40GB each,
external hard driverun llmollama hongkiat
https://www.hongkiat.com/blog/docker-llm-setup-guide/
How to Run LLM in Docker - Hongkiat
Large Language Models (LLMs) have changed how we build and use software. While cloud-based LLM APIs are great for convenience, there are plenty of reasons
run llmdockerhongkiat
https://simonwillison.net/2026/Mar/18/llm-in-a-flash/
Autoresearching Apple's "LLM in a Flash" to run Qwen 397B locally
Here's a fascinating piece of research by Dan Woods, who managed to get a custom version of Qwen3.5-397B-A17B running at 5.5+ tokens/second on a 48GB MacBook...
applellmflashrunqwen
https://simonwillison.net/2025/Feb/15/llm-mlx/
Run LLMs on macOS using llm-mlx and Apple’s MLX framework
llm-mlx is a brand new plugin for my LLM Python Library and CLI utility which builds on top of Apple’s excellent MLX array framework library and mlx-lm...
run llmsmacosusingmlxframework
https://www.bcdvideo.com/blog/what-hardware-is-needed-to-run-llms/
What Hardware Do You Need to Run an LLM? | BCD
Apr 7, 2026 - Interested in running a local large language model? Here's everything you should know about the various hardware you'll need for a high-performing LLM.
hardwareneedrunllmbcd
https://apxml.com/tools/vram-calculator
Can You Run This LLM? VRAM Calculator (Nvidia GPU and Apple Silicon)
Calculate the VRAM required to run any large language model.
nvidia gpuapple siliconrunllmvram
https://simonwillison.net/2024/Oct/27/llm-jq/
Run a prompt to generate and execute jq programs using llm-jq
llm-jq is a brand new plugin for LLM which lets you pipe JSON directly into the llm jq command along with a human-language description of how you’d like to...
programs usingrunpromptgenerateexecute
https://www.xda-developers.com/gemma-4-not-smartest-local-llm-but-reach-for/
Google's Gemma 4 isn't the smartest local LLM I've run, but it's the one I reach for most
Apr 15, 2026 - Google's newest Gemma 4 models are both powerful and useful.
gemma 4local llmgooglesmartestrun
https://novoserve.com/blog/how-to-run-an-llm-on-a-server-your-2026-llm-server-hardware-guide
How to Run an LLM on a Server: Your 2026 LLM Server Hardware Guide
Jan 21, 2026 - The rise of Large Language Models (LLMs) has been transformative, but how do you run an LLM on a server on your own?
2026 hardwarerunllmserverguide
https://simonwillison.net/2025/May/27/llm-tools/
Large Language Models can run tools in your terminal with LLM 0.26
LLM 0.26 is out with the biggest new feature since I started the project: support for tools. You can now use the LLM CLI tool—and Python library—to grant LLMs...
large language modelsllm 0runtoolsterminal
https://www.makeuseof.com/you-can-and-should-run-a-tiny-llm-on-your-android-phone/
You can (and should) run a tiny LLM on your Android phone
Feb 9, 2026 - Unused RAM is wasted RAM
android phoneruntinyllm
https://thenewstack.io/how-to-set-up-and-run-a-local-llm-with-ollama-and-llama-2/
How To Use Ollama: Set Up and Run a Local LLM With Llama 3
Jan 31, 2025 - Take a look at how to run an open source LLM locally, which allows you to run queries on your private data without any security concerns.
use ollamalocal llmsetrun3