Running a local LLM seems straightforward. You install Ollama, pull a model, and start chatting. But if you've spent any time actually using one, you've probably ...