OpenClaw is a personal AI assistant that connects your messaging apps to local AI coding agents, all running on your […]
ollama launch is a new command which sets up and runs coding tools like Claude Code, OpenCode, and Codex with […]
Generate images locally with Ollama on macOS. Windows and Linux support coming soon.
Ollama is now compatible with the Anthropic Messages API, making it possible to use tools like Claude Code with open […]
Open models can be used with OpenAI’s Codex CLI through Ollama. Codex can read, modify, and execute code in your […]
Ollama is partnering with OpenAI and ROOST (Robust Open Online Safety Tools) to bring the latest gpt-oss-safeguard reasoning models to […]
MiniMax M2 is now available on Ollama’s cloud. It’s a model built for coding and agentic workflows.
We ran performance tests on release day firmware and an updated Ollama version to see how Ollama performs.
GLM-4.6 and Qwen3-coder-480B are available on Ollama’s cloud service with easy integrations to the tools you are familiar with. Qwen3-Coder-30B […]
The latest NVIDIA DGX Spark is here! Ollama has partnered with NVIDIA to ensure it runs fast and efficiently out-of-the-box.
A new web search API is now available in Ollama. Ollama provides a generous free tier of web searches for […]
Cloud models are now in preview, letting you run larger models with fast, datacenter-grade hardware. You can keep using your […]
Ollama now includes a significantly improved model scheduling system, reducing crashes due to out of memory issues, maximizing GPU utilization […]
Ollama now supports streaming responses with tool calling. This enables all chat applications to stream content and also call tools […]
Ollama now has the ability to enable or disable thinking. This gives users the flexibility to choose the model’s thinking […]
Secure Minions is a secure protocol built by Stanford’s Hazy Research lab to allow encrypted local-remote communication.
With Ollama Python library version 0.4, functions can now be provided as tools. The library now also has full typing […]
Ollama now supports structured outputs making it possible to constrain a model’s output to a specific format defined by a […]
Avanika Narayan, Dan Biderman, and Sabri Eyuboglu from Christopher Ré’s Stanford Hazy Research lab, along with Avner May, Scott Linderman, […]
Ollama now supports new multimodal models with its new engine.
Continue enables you to easily create your own coding assistant directly inside Visual Studio Code and JetBrains with open-source LLMs.
Ollama now supports tool calling with popular models such as Llama 3.1. This enables a model to answer a given […]
Bespoke-Minicheck is a new grounded factuality checking model developed by Bespoke Labs that is now available in Ollama. It can […]
Embedding models are available in Ollama, making it easy to generate vector embeddings for use in search and retrieval augmented […]
Llama 3 is now available to run on Ollama. This model is the next generation of Meta’s state-of-the-art large language […]
Compared to Llama 2, Llama 3 feels much less censored. Meta has substantially lowered false refusal rates. Llama 3 will […]
At Google IO 2024, Google announced Ollama support in Firebase Genkit, a new open-source framework for developers to build, deploy […]
Ollama now has initial compatibility with the OpenAI Chat Completions API, making it possible to use existing tooling built for […]
Ollama is now available on Windows in preview, making it possible to pull, run and create large language models in […]
Ollama now supports AMD graphics cards in preview on Windows and Linux. All the features of Ollama can now be […]
The initial versions of the Ollama Python and JavaScript libraries are now available, making it easy to integrate your Python […]
New vision models are now available: LLaVA 1.6, in 7B, 13B and 34B parameter sizes. These models support higher resolution […]
Ollama can now run with Docker Desktop on the Mac, and run inside Docker containers with GPU acceleration on Linux.
Recreate one of the most popular LangChain use-cases with open source, locally running software – a chain that performs Retrieval-Augmented […]