by macOS26
Any AI, replaces Claude Code, Cursor, Cline, Open Claw. 17+ LLM providers (Claude, OpenAI, Gemini, Ollama, Zai, HF, Qwen) wired into a native Mac app that writes code, builds Xcode projects, bumps versions, manages git, Eautomates Safari, AppleScript 50 apps, extend w/ MCP Servers, run tasks from your iPhone via Messages, drive any Accessible app.
# Add to your Claude Code skills
git clone https://github.com/macOS26/AgentAgent/SDEFs/ plus every .app in /Applications, /System/Applications, ~/Applications are discovered at runtime — installing a new app extends what the agent can target with no code edit.cached_tokens is parsed from the response and shown in the LLM Usage panel. JSON request bodies use so byte-stable prefixes actually hit the provider's cache.No comments yet. Be the first to share your thoughts!
.sortedKeystieredCompact) — free, private, no API tokens consumed. Toggleable in the brain icon popover.LLMRegistry — all the previously-shipped fundamentals are still there.One app. Any AI. Total command over your Mac.
Agent! wires 17 LLM providers — Claude, GPT, Gemini, Grok, Mistral, DeepSeek, Qwen, Z.ai, BigModel, Hugging Face, Ollama (cloud and local), vLLM, LM Studio, Codestral, Mistral Vibe, and on-device Apple Intelligence — into a native macOS app that doesn't just talk about doing things. It does them.
Watch it read your codebase, fix the bug, build the Xcode project, and commit the diff while you make coffee. Tell it to open Safari and text you the price of flights to Tokyo. Say "Agent!" from across the room and have it run your test suite by voice. Text your Mac from iMessage and get a polished answer before you reach your car.
It edits files with surgical string-replace diffs — every change one-click undoable from a Time-Machine-style rollback. It drives any Mac app through the Accessibility API — no AppleScript required. It remembers your preferences across sessions. It spawns parallel sub-agents for work that fans out. It indexes entire codebases into a portable JSONL repo-map that any LLM can consume. It runs shell commands as you, or as root via a Launch Daemon you approve exactly once.
Bring your own API key. Run it fully local on Ollama, vLLM, or LM Studio. Or run it free, forever, on Apple Intelligence. No subscription. No telemetry. No vendor lock-in. Your keys, your machine, your data.
Download it. Say what you need. Watch it happen.
git clone https://github.com/toddbruss/Agent.git
cd Agent
Agent.xcodeproj in Xcode.Agent target../build.sh # Debug build
./build.sh Release # Release build
build/DerivedData/Build/Products/Debug/Agent!.appopen "build/DerivedData/Build/Products/Debug/Agent!.app"⚠️ Without a developer account the app is ad-hoc signed. The Launch Agent/Daemon helpers won't register (SMAppService needs a team ID), but the LLM loop, all tools, accessibility, AppleScript, shell, and MCP all work.
💡 Cheapest cloud path? GLM-5.1 (the latest) is now available on all four of the cheap cloud providers — Ollama, Hugging Face, Z.ai, and BigModel. Pennies per million tokens vs Claude/GPT pricing. Pick whichever you already have an account with; pricing is competitive across all of them.
💡 Z.ai is still the recommended starting point if you don't have an account anywhere yet — fastest signup, GLM-5.1 is the default model, no infrastructure to provision.
⚠️ Running GLM locally? Only GLM-4.7-Turbo (32B) runs well on consumer hardware — M2/M3/M4 Mac with 64-128GB unified memory via Ollama. GLM-5 (744B MoE) and GLM-5.1 (754B MoE) are too large to run locally (~1.6TB full weight) — use them via Z.ai, BigModel, Hugging Face cloud, or Ollama cloud.
"Play my Workout playlist in Music" "Build the Xcode project and fix any errors" "Take a photo with Photo Booth" "Send an iMessage to Mom saying I'll be home at 6" "Open Safari and search for flights to Tokyo" "Refactor this class into smaller files" "What calendar events do I have today?"
Just type what you want. Agent! figures out how and makes it happen.
Built-in autonomous task loop that reasons, executes, and self-corrects. Agent! doesn't just run code; it observes the results, debugs errors, and iterates until the task is complete.
Full coding environment built in. Reads codebases, edits files with precision, runs shell commands, builds Xcode projects, manages git, and auto-enables coding mode to focus the AI on development tools. Replaces Claude Code, Cursor, and Cline -- no terminal, no IDE plugins, no monthly fee. Features Time Machine-style backups for every file change, letting you revert any edit instantly.
Automatically detects and uses available tools (Xcode, Playwright, Shell, etc.) based on your prompt. No manual configuration required for core tools.
Securely runs root-level commands via a dedicated macOS Launch Daemon. The user approves the daemon once, then the agent can execute commands autonomously via XPC.
Control any Mac app through the Accessibility API. Click buttons, type into fields, navigate menus, scroll, drag -- all programmatically. Powered by AXorcist for reliable, fuzzy-matched element finding.
The provider picker (LLM Settings, toolbar button #7) shows 16 providers; Apple Intelligence is reached via the separate brain icon (#8). Source of truth: AgentTools.APIProvider.
| Provider | API key | Best for | |---|---|---| | Claude (Anthropic) | Paid | Long autonomous tasks, complex reasoning, prompt caching | | OpenAI | Paid | General purpose, tool calling, vision | | Google Gemini | Paid (free tier) | Long context, vision, fast | | Grok (xAI) | Paid | Real-time info | | Mistral | Paid | Open-weight cloud, fast tool calling | | Codestral (Mistral) | Paid | Code-specialized Mistral | | Mistral Vibe | Paid | Mistral's chat/agent product | | DeepSeek | Cheap | Budget cloud, strong coding, prompt cache hit reporting | | Hugging Face | Varies | Open-source models hosted serverless or on dedicated endpoints | | Z.ai | Cheap | GLM-5.1 via API — recommended starting point | | BigModel (Zhipu) | Cheap | GLM family via Zhipu's API | | Qwen (Alibaba) | Cheap | Qwen 2.5 / 3 via Dashscope | | Ollama (cloud) | Free tier | Run open models via Ollama's hosted endpoint | | Local Ollama | Free + hardware | Self-hosted Ollama daemon — fully offline, no account | | vLLM | Free + hardware | Self-hosted vLLM server with prefix caching | | LM Studio | Free + hardware | Self-hosted, easiest GUI for local models | | Apple Intelligence | Free, on-device | Triage, summary, accessibility intent (via brain icon, not the provider picker) |
💡 Self-hosted "free" providers (Local Ollama, vLLM, LM Studio) are only free in the API-fee sense. Running a 30B+ model with usable speed needs an M2/M3/M4 Ultra Mac Studio (64-128GB unified