Ollama's preview release on Windows allows users to pull, run, and create large language models with a native experience that includes built-in GPU acceleration, full access to the model library, and compatibility with the Ollama API and OpenAI. The software utilizes NVIDIA GPUs and modern CPU instruction sets for hardware acceleration, eliminating the need for configuration or virtualization. Users can run vision models, such as LLaVA 1.6, by simply dragging and dropping images into the interface. The always-on Ollama API runs in the background, enabling easy connectivity for tools and applications, and supports existing OpenAI-compatible tools, allowing integration with local models. The setup process involves downloading the installer, and users are encouraged to provide feedback or report issues through a Discord server.