Member-only story
Running Ollama on Windows Made Easy
Table of Contents
What’s New in Ollama on Windows?
Step-by-Step Guide to Running Ollama on Windows
Ollama, the versatile platform for running large language models (LLMs) locally, is now available on Windows. This update empowers Windows users to pull, run, and create LLMs with a seamless native experience. Packed with features like GPU acceleration, access to an extensive model library, and OpenAI-compatible APIs, Ollama on Windows is designed to deliver a robust and efficient AI development environment.
What’s New in Ollama on Windows?
1. Native Windows Experience
The Windows preview brings Ollama’s capabilities to a new audience, offering:
GPU Acceleration: Built-in support for NVIDIA GPUs and modern CPU instruction sets like AVX and AVX2 ensures faster model performance. No configuration or virtualization is required!
Full Model Library Access: From language models like Llama 2 to vision models like LLaVA 1.6, the entire Ollama library is now accessible on Windows. Vision models even allow drag-and-drop image inputs in the terminal during runtime.
Always-On API: Ollama’s API runs automatically in the background on http://localhost:11434, allowing tools and applications to connect seamlessly.
2. OpenAI Compatibility
Ollama on Windows supports the same OpenAI-compatible API as its macOS counterpart. This means you can integrate Ollama with existing OpenAI-compatible tooling and workflows for local model execution.
Step-by-Step Guide to Running Ollama on Windows
1. Get Started
Download Ollama on Windows
Visit Ollama’s website and download the Windows preview installer.