Serve local large language models natively on Apple Silicon with OpenAI/Ollama-compatible APIs, supporting tool calling, a plugin ecosystem, and a menu bar chat UI.
Sponsored
Osaurus is an all-in-one local LLM server specifically engineered for macOS with Apple Silicon, leveraging the MLX runtime for optimized inference. It provides OpenAI and Ollama compatible APIs, allowing seamless integration with existing AI tools and SDKs, including advanced features like tool calling. Beyond a robust server, Osaurus acts as a full Model Context Protocol (MCP) server, enabling AI agents to access an extensible plugin ecosystem for enhanced capabilities like browser automation, file system interaction, and web search. It also features a convenient menu bar chat UI and supports Apple Foundation Models on compatible macOS versions.
Key Features
01Extensible plugin system with official tools (browser, filesystem, git, search)
02Native Apple Silicon LLM server with MLX optimization
03Built-in Menu Bar Chat UI and Model Manager
04Full MCP (Model Context Protocol) server capabilities
05OpenAI & Ollama compatible APIs with streaming and tool calling
061,599 GitHub stars
Use Cases
01Run various open-source LLMs locally on macOS for privacy and performance
02Integrate local LLMs and their tools with AI agents and development environments via MCP
03Develop custom AI-powered applications leveraging a local OpenAI-compatible API