Forget Docker—this episode shows how to run Ollama natively on Windows, with full GPU acceleration.
What’s Inside:
- How to set up Ollama on Windows and configure it to use GPUs
- Comparison of performance: CPU vs GPU
- Tips for installing LLMs like LLaMA 3 on dev machines
- Demo: Using a GPU to serve prompts faster from SQL Server