🐧 PenguinPulse

Linux Graphics & Gaming News

Articles tagged: ollama

← Back to all articles

Ollama v0.23.1 Accelerates Gemma 4 MLX Inference with MTP Speculative Decoding

Ollama has released version 0.23.1, introducing Gemma 4 Multi-token Processing (MTP) speculative decoding for its MLX runner. This update, primarily benefiting macOS users with Apple Silicon, can prov...

Read more →

Ollama 0.23.0 Adds Claude Desktop Integration, Server-Driven Model Recommendations

Ollama v0.23.0 was released today, introducing official support for the Claude Desktop application. This integration allows users to run both Claude Cowork and Claude Code directly within the Claude D...

Read more →

Ollama v0.22.1-rc1 Adds NVIDIA TensorRT Import, Enhanced Model Batching

Ollama has released v0.22.1-rc1, bringing notable enhancements for local AI model execution. Key updates include the addition of NVIDIA TensorRT Model Optimizer import support, which allows for levera...

Read more →

Ollama v0.21.1 Adds Kimi CLI, Enhances MLX Runner Performance and Features

Ollama released version 0.21.1 recently, introducing the Kimi CLI for local AI model execution. Users can now launch the Kimi CLI via Ollama, which "excels at long horizon agentic execution tasks thro...

Read more →

Ollama v0.20.6 Delivers Gemma 4 Tool Calling Updates, Parallel Streaming Fixes

Ollama released version 0.20.6 today, focusing on enhancements for local AI deployments. The update notably improves Gemma 4 tool calling capabilities, incorporating Google's latest post-launch fixes ...

Read more →

Ollama v0.21.0 Integrates Nous Research's Hermes Agent for Adaptive AI Workflows

Ollama v0.21.0 was released today, introducing the Hermes Agent for enhanced local AI workflows. This integration allows users to deploy Nous Research's self-improving AI agent via the 'ollama launch ...

Read more →

Ollama v0.20.7 Updates ROCm to 7.2.1 for Linux, Addresses Gemma Model Quality

Ollama, a framework designed for running large language models locally, recently released version 0.20.7. This update brings key improvements for Linux users, especially those utilizing AMD GPUs for t...

Read more →

Ollama v0.20.8-rc0 Updates ROCm to 7.2.1 on Linux

Ollama announced its v0.20.8-rc0 pre-release candidate recently, introducing several updates with a key focus on improved hardware support. For Linux users, this version notably updates its AMD ROCm i...

Read more →

Ollama v0.20.5 Debuts OpenClaw Multi-Channel Integration, Gemma 4 Flash Attention

Ollama v0.20.5 was released on April 9, 2026, introducing significant new features for local AI model interaction and performance. The update includes the OpenClaw channel setup, which allows users to...

Read more →

Ollama v0.20.0 Introduces Gemma 4 Models; v0.20.4-rc Improves MLX & Flash Attention

Ollama recently released version 0.20.0, which introduces support for the Gemma 4 series of large language models. This includes "Effective 2B (E2B)", "Effective 4B (E4B)", "26B (Mixture of Experts mo...

Read more →

Ollama v0.19.0 Enhances Local AI with Web Search Plugin, KV Cache Efficiency

Ollama, the local AI framework, released version 0.19.0 today, March 30, 2026. This update introduces a new web search plugin for its 'ollama launch pi' command, enabling direct web search capabilitie...

Read more →

Ollama v0.18.3 Integrates Local AI Models Directly with VS Code via Copilot

Ollama released version 0.18.3 yesterday, introducing direct integration with Microsoft Visual Studio Code through GitHub Copilot. This update allows developers to select and utilize any local or clou...

Read more →

Ollama v0.18.2 Boosts Local Claude Code Speed, Refines OpenClaw Integration

Ollama has released version 0.18.2 today, introducing performance improvements and fixes for its integration with OpenClaw. A key update in this release is the enhanced speed of local AI code generati...

Read more →

Ollama v0.18.1 Integrates Web Search, Fetch for OpenClaw Models

Ollama v0.18.1 was released, introducing new web search and web fetch plugins specifically for OpenClaw. This update allows Ollama's language models, whether local or cloud-based, to search the web fo...

Read more →

Ollama 0.18.0 Boosts OpenClaw Performance, Integrates as Provider

Ollama version 0.18.0 was released today, bringing significant performance improvements and new integration capabilities for AI model usage. The update enhances the speed of OpenClaw and Ollama's clou...

Read more →

Ollama v0.17.5 Adds Qwen 3.5 Models, Fixes GPU/CPU Split & MLX Crashes

Ollama v0.17.5 was released today, introducing support for the Qwen 3.5 small model series, now available in 0.8B, 2B, 4B, and 9B parameter sizes. This update addresses several critical issues, includ...

Read more →

Ollama v0.16.0 Introduces New GLM-5, MiniMax-M2.5 LLMs and CLI Tools

Ollama released version 0.16.0 yesterday, enhancing its local large language model ecosystem with new models and a simplified command-line interface. The update introduces GLM-5, a 744 billion paramet...

Read more →

Ollama 0.15.0 Adds 'ollama launch' for AI Code Model Integration

Ollama 0.15.0 was released today, introducing the new "ollama launch" command. This command allows users to directly integrate Ollama's models with various large language models, including Claude Code...

Read more →

Ollama 0.14.3 Adds Z-Image Turbo Text-to-Image and GLM-4.7-Flash LLM

Ollama released version 0.14.3 three days ago, introducing several new large language models and text-to-image capabilities. Key additions include Z-Image Turbo, a 6 billion parameter text-to-image mo...

Read more →

Ollama 0.14.2 Introduces TranslateGemma Models for 55-Language Translation

Ollama released version 0.14.2 today, January 19, 2026, a minor update following the experimental image generation models in 0.14.1. The highlight of this release is the introduction of "TranslateGemm...

Read more →

Ollama 0.14.1 Introduces Experimental Image Generation Models for Linux (CUDA)

Ollama v0.14.1 was released today, January 18, 2026, bringing experimental support for image generation models. This new functionality is available for Linux systems configured with CUDA-compatible GP...

Read more →

Ollama v0.13.2-rc2 Boosts Vision Model AI Performance, Fixes Multi-GPU CUDA

Ollama has recently released version 0.13.2-rc2, bringing several updates focused on AI model performance and GPU compatibility. A significant change is the default enablement of flash attention for v...

Read more →

Ollama v0.13.1-rc0 Improves CUDA VRAM Discovery, Adds Cogito-v2.1 Tool Calling

Ollama recently released version 0.13.1-rc0, introducing several updates for local large language model development. This release addresses specific issues with CUDA VRAM discovery, which should lead ...

Read more →