by digitalsamba
AI-native video production toolkit for Claude Code
# Add to your Claude Code skills
git clone https://github.com/digitalsamba/claude-code-video-toolkitAn AI-native video production workspace for Claude Code. Skills, commands, templates, and tools that give Claude Code everything it needs to help you create professional videos — from concept to final render.
git clone https://github.com/digitalsamba/claude-code-video-toolkit.git
cd claude-code-video-toolkit
python3 -m pip install -r tools/requirements.txt # Optional: AI voiceover, image gen, music, moviepy examples
claude # Open Claude Code in the toolkit
Then in Claude Code:
/setup # Configure cloud GPU, storage, voice (~5 min, mostly free)
/video # Create your first video
That's it. /setup walks you through everything interactively — cloud GPU provider, file transfer, voice config. /video creates a project from a template and guides you through the whole workflow.
What's free: The toolkit leans heavily on open-source AI models — voiceovers (Qwen3-TTS), image generation (FLUX.2), music (ACE-Step), and more. You deploy them to your own cloud GPU account and run them at cost. Cloudflare R2 has a generous free tier (10GB, zero egress), and Modal gives $30/month free compute on the Starter plan — more than enough for a few 5-minute videos a month.
Requirements: Node.js 18+ and Claude Code. Python 3.9+ recommended for AI tools. FFmpeg optional.
No comments yet. Be the first to share your thoughts!
Want to skip setup and just render something?
cd examples/hello-world && npm install && npm run renderNo API keys needed — outputs an MP4 immediately.
I've spent months painstakingly putting this toolkit together and plan to keep iterating on it. AI makes things easier, but hard work still has huge value. Every video I create is a chance for improvement — every skill, template, tool, and workflow here has been refined through that cycle. It would be wonderful if others wanted to get involved with that: use it, refine it, and feed back into the repo via an issue or PR what you learn.
My own use case is fairly specific: creating sprint review videos for the AI mobile development arm of Digital Samba. But the idea behind this project is a reusable toolkit for using Claude Code to autonomously generate any kind of "explainer" style video — product demos, walkthroughs, presentations, whatever you need. Autonomous video creation is a lofty ideal for such a subjective field, but we can try :)
What makes this work is that Claude Code is fantastically resourceful and flexible — give it the framing and tooling that this toolkit provides and it will adapt it to create templates and videos based on your prompting. The skills, templates, and tools here are building blocks. Claude Code is the builder. You are the director, editor, and designer.
If you're getting started, run
/setupthen/videoand let Claude Code guide you. Or start with/templateto create a template for your own use case.Cloud GPU — I recommend Modal for running the toolkit's AI tools. The Starter plan gives you $30/month free compute, which is more than enough. RunPod is also supported as an alternative. Run
/setupto deploy the tools you need.My motto: Be brave. Experiment. And please share any videos you create or ideas you have back with the project — it helps me keep improving this toolkit for everyone.
Claude Code has deep knowledge in:
| Skill | Description | |-------|-------------| | remotion | React-based video framework — compositions, animations, rendering | | elevenlabs | AI audio — text-to-speech, voice cloning, music, sound effects | | ffmpeg | Media processing — format conversion, compression, resizing | | playwright-recording | Browser automation — record demos as video | | frontend-design | Visual design refinement for distinctive, production-grade aesthetics | | qwen-edit | AI image editing — prompting patterns and best practices | | acestep | AI music generation — prompts, lyrics, scene presets, video integration | | ltx2 | AI video generation — text-to-video, image-to-video clips, prompting guide | | moviepy | Python video composition — overlay text on LTX-2/SadTalker output, build.py-style projects | | runpod | Cloud GPU — setup, Docker images, endpoint management, costs |
| Command | Description |
|---------|-------------|
| /setup | First-time setup — cloud GPU, file transfer, voice, prerequisites |
| /video | Video projects — list, resume, or create new |
| /scene-review | Scene-by-scene review in Remotion Studio |
| /design | Focused design refinement session for a scene |
| /brand | Brand profiles — list, edit, or create new |
| /template | List available templates or create new ones |
| /skills | List installed skills or create new ones |
| /contribute | Share improvements — issues, PRs, examples |
| /record-demo | Record browser interactions with Playwright |
| /generate-voiceover | Generate AI voiceover from a script |
| /redub | Redub existing video with a different voice |
| /voice-clone | Record, test, and save a cloned voice to a brand |
| /versions | Check dependency versions and toolkit updates |
Note: After creating or modifying commands/skills, restart Claude Code to load changes.
Pre-built video structures in templates/:
See examples/ for finished projects you can learn from (oldest first, showing toolkit evolution):
| Date | Demo | Description | |------|------|-------------| | 2025-12-05 | sprint-review-cho-oyu | iOS sprint review with demos | | 2025-12-10 | digital-samba-skill-demo | Product demo showcasing Claude Code skill | | 2026-01-22 | ds-remote-mcp | Remote MCP server demo (the jazz background music is a joke) | | 2026-01-25 | schlumbergera | Android sprint review video | | 2026-02-23 | cortina | Mobile platforms sprint review | | 2026-03-15 | the-space-between | AI-generated video essay — flux2 avatar, Qwen3-TTS voice, SadTalker animation | | 2026-04-08 | q2-townhall-longarm-ad | Super Bowl-style launch ad with dramatic Qwen3-TTS announcer and LTX-2 animated Lugh cameo | | 2026-04-08 | q2-townhall-stars | GitHub star history time-lapse with animated chart and deadpan-to-excited commentary |
The toolkit includes a transitions library for scene-to-scene effects:
| Transition | Description |
|------------|-------------|
| glitch() | Digital distortion with RGB shift |
| rgbSplit() | Chromatic aberration effect |
| zoomBlur() | Radial motion blur |
| lightLeak() | Cinematic lens flare |
| clockWipe() | Radial sweep reveal |
| pixelate() | Digital mosaic dissolution |
| checkerboard() | Grid-based reveal (9 patterns) |
Plus official Remotion transitions: slide(), fade(), wipe(), flip()
Preview all transitions:
cd showcase/transitions && npm install && npm run studio
See lib/transitions/README.md for full documentation.
Define visual identity in brands/. When you create a project with /video, the brand's colors, fonts, and styling are automatically applied.
brands/my-brand/
├── brand.json # Colors, fonts, typography
├── voice.json # ElevenLabs voice settings
└── assets/ # Logo, backgrounds
Included brands: default, digital-samba
Create your own with /brand.
Video projects are tracked through a multi-session lifecycle:
planning → assets → review → audio → editing → rendering → complete
Each project has a project.json that tracks:
The system automatically reconciles intent (what you planned) with reality (what files exist), and generates a CLAUDE.md per project for instant context when resuming.
See lib/project/README.md for schema details, scene status tracking, and filesystem reconciliation logic.
Audio, video, and image tools in tools/:
# Generate voiceover (ElevenLabs)
python tools/voiceover.py --script script.md --output voiceover.mp3
# Generate voiceover (Qwen3-TTS — self-hosted, cheaper alternative)
python tools/voiceover.py --provider qwen3 --speaker Ryan --scene-dir public/audio/scenes --json
python tools/qwen3_tts.py --text "Hello world" --tone warm --output hello.mp3
# Generate background music (ElevenLabs)
python tools/music.py --prompt "Upbeat corporate" --duration 120 --outp