Local LLMs in VS Code with GitHub Copilot Chat and Ollama

How to run fully local language models in VS Code’s Copilot Chat using Ollama — including a structured approach to maintaining separate planning and coding models per machine.

April 23, 2026 · 4 min · 830 words · Felix Itzenplitz