Today, we’re previewing the fastest way to run Ollama on Apple silicon, powered by MLX, Apple’s machine learning framework.
Setup OpenClaw in under two minutes with a single Ollama command.
OpenClaw is a personal AI assistant that connects your messaging apps to local AI coding agents, all running on your […]
ollama launch is a new command which sets up and runs coding tools like Claude Code, OpenCode, and Codex with […]
Generate images locally with Ollama on macOS. Windows and Linux support coming soon.
Ollama is now compatible with the Anthropic Messages API, making it possible to use tools like Claude Code with open […]
Open models can be used with OpenAI’s Codex CLI through Ollama. Codex can read, modify, and execute code in your […]
Ollama is partnering with OpenAI and ROOST (Robust Open Online Safety Tools) to bring the latest gpt-oss-safeguard reasoning models to […]
MiniMax M2 is now available on Ollama’s cloud. It’s a model built for coding and agentic workflows.
We ran performance tests on release day firmware and an updated Ollama version to see how Ollama performs.
GLM-4.6 and Qwen3-coder-480B are available on Ollama’s cloud service with easy integrations to the tools you are familiar with. Qwen3-Coder-30B […]
The latest NVIDIA DGX Spark is here! Ollama has partnered with NVIDIA to ensure it runs fast and efficiently out-of-the-box.
A new web search API is now available in Ollama. Ollama provides a generous free tier of web searches for […]
Ollama now includes a significantly improved model scheduling system, reducing crashes due to out of memory issues, maximizing GPU utilization […]
Cloud models are now in preview, letting you run larger models with fast, datacenter-grade hardware. You can keep using your […]
Secure Minions is a secure protocol built by Stanford’s Hazy Research lab to allow encrypted local-remote communication.