Accessibility review agents for Claude Code, GitHub Copilot, and Claude Desktop. Eleven specialists that enforce WCAG 2.2 AA compliance so AI coding tools stop generating inaccessible code.
# Add to your Claude Code skills
git clone https://github.com/Community-Access/accessibility-agentsAI and automated tools are not perfect. They miss things, make mistakes, and cannot replace testing with real screen readers and assistive technology. Always verify with VoiceOver, NVDA, JAWS, and keyboard-only navigation. This tooling is a helpful starting point, not a substitute for real accessibility testing.
A community-driven open-source project automating accessibility, efficiency, and productivity through AI-based agents, skills, custom instructions, and prompts.
A sincere thanks goes out to Taylor Arndt and Jeff Bishop for leading the charge in building this community project. It started because LLMs consistently forget accessibility - skills get ignored, instructions drift out of context, ARIA gets misused, focus management gets skipped, color contrast fails silently. They got tired of fighting it and built an agent team that will not let it slide. Now we want to make more magic together.
If you care about making software accessible to blind and low vision users, please consider . Every improvement to these agents helps developers ship more inclusive software for the people who need it most.
No comments yet. Be the first to share your thoughts!
AI coding tools generate inaccessible code by default. They forget ARIA rules, skip keyboard navigation, ignore contrast ratios, and produce modals that trap screen reader users. Even with skills and CLAUDE.md instructions, accessibility context gets deprioritized or dropped entirely.
Accessibility Agents provides thirty-five specialized agents across two teams and six platforms:
All agents run on: