Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
Created by
brew bumpCreated with
brew bump-formula-pr.Details
release notes
/v1/messagesendpoint for seamless drop-in replacement of Claude.🚀 New Features & Major Enhancements
🤖 Open Responses API: Build Smarter, Autonomous Agents
LocalAI now supports the OpenAI Responses API, enabling powerful agentic workflows locally.
response_id— resume and manage long-running agent sessions.Our support passes all the official acceptance tests:
🧠 Anthropic Messages API: Clone Claude Locally
LocalAI now fully supports the Anthropic
messagesAPI.https://api.localai.host/v1/messagesas a drop-in replacement for Claude.anthropic-sdk-go, LangChain, and other tooling.🎥 Video Generation: From Text to Video in the Web UI
diffuserswith full compatibility.⚙️ Unified GPU Backends: Acceleration Works Out of the Box
A major architectural upgrade: GPU libraries (CUDA, ROCm, Vulkan) are now packaged inside backend containers.
Note: this is experimental, please help us by filing an issue if something doesn't work!
🧩 Tool Streaming & Advanced Parsing
Enhance your agent workflows with richer tool interaction.
input_json_delta).<function>...</function>) are now properly parsed alongside text.🌐 System-Aware Backend Gallery: Only Compatible Backends Show
The backend gallery now shows only backends your system can run.
🎤 New TTS Backends: Pocket-TTS
Add expressive voice generation to your apps with Pocket-TTS.
🔍 Request Tracing: Debug Your Agents
Trace requests and responses in memory — great for fine-tuning and agent debugging.
GET /api/v1/trace.🪄 New 'Reasoning' Field: Extract Thinking Steps
LocalAI now automatically detects and extracts thinking tags from model output.
🚀 Moonshine Backend: Faster Transcription for Low-End Devices
Add Moonshine, an ONNX-based transcription engine, for fast, lightweight speech-to-text.
🛠️ Fixes & Stability Improvements
🔧 Prevent BMI2 Crashes on AVX-Only CPUs
Fixed crashes on older Intel CPUs (Ivy Bridge, Sandy Bridge) that lack BMI2 instructions.
llama-cpp-fallback(SSE2 only).EOFerrors during model warmup.📊 Fix Swapped VRAM Usage on AMD GPUs
Correctly parses
rocm-smioutput: used and total VRAM are now displayed correctly.HIP_VISIBLE_DEVICESproperly (e.g., when using only discrete GPU).🚀 The Complete Local Stack for Privacy-First AI
LocalAI
The free, Open Source OpenAI alternative. Drop-in replacement REST API compatible with OpenAI specifications for local AI inferencing. No GPU required.
Link: https://github.com/mudler/LocalAI
LocalAGI
Local AI agent management platform. Drop-in replacement for OpenAI's Responses API, supercharged with advanced agentic capabilities and a no-code UI.
Link: https://github.com/mudler/LocalAGI
LocalRecall
RESTful API and knowledge base management system providing persistent memory and storage capabilities for AI agents. Works alongside LocalAI and LocalAGI.
Link: https://github.com/mudler/LocalRecall
❤️ Thank You
LocalAI is a true FOSS movement — built by contributors, powered by community.
If you believe in privacy-first AI:
Your support keeps this stack alive.
✅ Full Changelog
📋 Click to expand full changelog
What's Changed
Bug fixes :bug:
Exciting New Features 🎉
🧠 Models
📖 Documentation and examples
👒 Dependencies
Other Changes
6114e692136bea917dc88a5eb2e532c3d133d963by @localai-bot in chore: ⬆️ Update ggml-org/whisper.cpp to6114e692136bea917dc88a5eb2e532c3d133d963mudler/LocalAI#7717c18428423018ed214c004e6ecaedb0cbdda06805by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toc18428423018ed214c004e6ecaedb0cbdda06805mudler/LocalAI#771885c40c9b02941ebf1add1469af75f1796d513ef4by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to85c40c9b02941ebf1add1469af75f1796d513ef4mudler/LocalAI#77317ac8902133da6eb390c4d8368a7d252279123942by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to7ac8902133da6eb390c4d8368a7d252279123942mudler/LocalAI#7740a4bf35889eda36d3597cd0f8f333f5b8a2fcaefcby @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toa4bf35889eda36d3597cd0f8f333f5b8a2fcaefcmudler/LocalAI#77514ffc47cb2001e7d523f9ff525335bbe34b1a2858by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to4ffc47cb2001e7d523f9ff525335bbe34b1a2858mudler/LocalAI#7760c9a3b40d6578f2381a1373d10249403d58c3c5bdby @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toc9a3b40d6578f2381a1373d10249403d58c3c5bdmudler/LocalAI#77780f89d2ecf14270f45f43c442e90ae433fd82dab1by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to0f89d2ecf14270f45f43c442e90ae433fd82dab1mudler/LocalAI#7795e9898ddfb908ffaa7026c66852a023889a5a7202by @localai-bot in chore: ⬆️ Update ggml-org/whisper.cpp toe9898ddfb908ffaa7026c66852a023889a5a7202mudler/LocalAI#781013814eb370d2f0b70e1830cc577b6155b17aee47by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to13814eb370d2f0b70e1830cc577b6155b17aee47mudler/LocalAI#7809ced765be44ce173c374f295b3c6f4175f8fd109bby @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toced765be44ce173c374f295b3c6f4175f8fd109bmudler/LocalAI#7822706e3f93a60109a40f1224eaf4af0d59caa7c3aeby @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to706e3f93a60109a40f1224eaf4af0d59caa7c3aemudler/LocalAI#7836e57f52334b2e8436a94f7e332462dfc63a08f995by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toe57f52334b2e8436a94f7e332462dfc63a08f995mudler/LocalAI#7848b90b1ee9cf84ea48b478c674dd2ec6a33fd504d6by @localai-bot in chore: ⬆️ Update leejet/stable-diffusion.cpp tob90b1ee9cf84ea48b478c674dd2ec6a33fd504d6mudler/LocalAI#78624974bf53cf14073c7b66e1151348156aabd42cb8by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to4974bf53cf14073c7b66e1151348156aabd42cb8mudler/LocalAI#7861c5602a676caff5fe5a9f3b76b2bc614faf5121a5by @localai-bot in chore: ⬆️ Update leejet/stable-diffusion.cpp toc5602a676caff5fe5a9f3b76b2bc614faf5121a5mudler/LocalAI#7880679bdb53dbcbfb3e42685f50c7ff367949fd4d48by @localai-bot in chore: ⬆️ Update ggml-org/whisper.cpp to679bdb53dbcbfb3e42685f50c7ff367949fd4d48mudler/LocalAI#7879e443fbcfa51a8a27b15f949397ab94b5e87b2450by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toe443fbcfa51a8a27b15f949397ab94b5e87b2450mudler/LocalAI#7881ccbc84a5374bab7a01f68b129411772ddd8e7c79by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toccbc84a5374bab7a01f68b129411772ddd8e7c79mudler/LocalAI#78949be0b91927dfa4007d053df72dea7302990226bbby @localai-bot in chore: ⬆️ Update leejet/stable-diffusion.cpp to9be0b91927dfa4007d053df72dea7302990226bbmudler/LocalAI#7895ae9f8df77882716b1702df2bed8919499e64cc28by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toae9f8df77882716b1702df2bed8919499e64cc28mudler/LocalAI#79150e52afc6513cc2dea9a1a017afc4a008d5acf2b0by @localai-bot in chore: ⬆️ Update leejet/stable-diffusion.cpp to0e52afc6513cc2dea9a1a017afc4a008d5acf2b0mudler/LocalAI#7930593da7fa49503b68f9f01700be9f508f1e528992by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to593da7fa49503b68f9f01700be9f508f1e528992mudler/LocalAI#7946b1377188784f9aea26b8abde56d4aee8c733eec7by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp tob1377188784f9aea26b8abde56d4aee8c733eec7mudler/LocalAI#79650c3b7a9efebc73d206421c99b7eb6b6716231322by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to0c3b7a9efebc73d206421c99b7eb6b6716231322mudler/LocalAI#7978885e62ea822e674c6837a8225d2d75f021b97a6aby @localai-bot in chore: ⬆️ Update leejet/stable-diffusion.cpp to885e62ea822e674c6837a8225d2d75f021b97a6amudler/LocalAI#7979bcf7546160982f56bc290d2e538544bbc0772f63by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp tobcf7546160982f56bc290d2e538544bbc0772f63mudler/LocalAI#79917010bb4dff7bd55b03d35ef9772142c21699eba9by @localai-bot in chore: ⬆️ Update leejet/stable-diffusion.cpp to7010bb4dff7bd55b03d35ef9772142c21699eba9mudler/LocalAI#8013a96310871a3b294f026c3bcad4e715d17b5905feby @localai-bot in chore: ⬆️ Update ggml-org/whisper.cpp toa96310871a3b294f026c3bcad4e715d17b5905femudler/LocalAI#8014e4832e3ae4d58ac0ecbdbf4ae055424d6e628c9fby @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp toe4832e3ae4d58ac0ecbdbf4ae055424d6e628c9fmudler/LocalAI#801547af2fb70f7e4ee1ba40c8bed513760fdfe7a704by @localai-bot in chore: ⬆️ Update ggml-org/whisper.cpp to47af2fb70f7e4ee1ba40c8bed513760fdfe7a704mudler/LocalAI#8039d98b548120eecf98f0f6eaa1ba7e29b3afda9f2eby @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp tod98b548120eecf98f0f6eaa1ba7e29b3afda9f2emudler/LocalAI#80402eeeba56e9edd762b4b38467bab96c2517163158by @localai-bot in chore: ⬆️ Update ggml-org/whisper.cpp to2eeeba56e9edd762b4b38467bab96c2517163158mudler/LocalAI#8052785a71008573e2d84728fb0ba9e851d72d3f8fabby @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to785a71008573e2d84728fb0ba9e851d72d3f8fabmudler/LocalAI#8053388ce822415f24c60fcf164a321455f1e008cafbby @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to388ce822415f24c60fcf164a321455f1e008cafbmudler/LocalAI#8073f53dc74843e97f19f94a79241357f74ad5b691a6by @localai-bot in chore: ⬆️ Update ggml-org/whisper.cpp tof53dc74843e97f19f94a79241357f74ad5b691a6mudler/LocalAI#80742fbde785bc106ae1c4102b0e82b9b41d9c466579by @localai-bot in chore: ⬆️ Update ggml-org/llama.cpp to2fbde785bc106ae1c4102b0e82b9b41d9c466579mudler/LocalAI#80879565c7f6bd5fcff124c589147b2621244f2c4aa1by @localai-bot in chore: ⬆️ Update leejet/stable-diffusion.cpp to9565c7f6bd5fcff124c589147b2621244f2c4aa1mudler/LocalAI#8086New Contributors
Full Changelog: mudler/LocalAI@v3.9.0...v3.10.0
View the full release notes at https://github.com/mudler/LocalAI/releases/tag/v3.10.0.