by SamurAIGPT
Multi-modal Generative Media Skills for AI Agents (Claude Code, Cursor, Gemini CLI). High-quality image, video, and audio generation powered by muapi.ai.
# Add to your Claude Code skills
git clone https://github.com/SamurAIGPT/Generative-Media-SkillsThe Ultimate Multimodal Toolset for Claude Code, Cursor, and Gemini CLI. A high-performance, schema-driven architecture for AI agents to generate, edit, and display professional-grade images, videos, and audio — powered by the muapi-cli.
🚀 Get Started | 🎨 Expert Library | ⚙️ Core Primitives | 🤖 MCP Server | 📖 Reference
--jq filtering for seamless agentic pipelines.muapi-cli — no curl, no JSON parsing, no boilerplate.--view flag to automatically download and open generated media in your system viewer.muapi mcp serve to expose all 19 tools directly to Claude Desktop, Cursor, or any MCP-compatible agent.No comments yet. Be the first to share your thoughts!
This repository uses a Core/Library split to ensure efficiency and high-signal discovery for LLMs:
/core)Thin wrappers around muapi-cli for raw API access.
core/media/ — File uploadcore/edit/ — Image editing (prompt-based)core/platform/ — Setup, auth & result polling/library)High-value skills that translate creative intent into technical directives.
/library/motion/cinema-director/) — Technical film direction & cinematography./library/visual/nano-banana/) — Reasoning-driven image generation (Gemini 3 Style)./library/visual/ui-design/) — High-fidelity mobile/web mockups (Atomic Design)./library/visual/logo-creator/) — Minimalist vector branding (Geometric Primitives)./library/motion/seedance-2/) — Director-level cinematic video generation with text-to-video, image-to-video, and video extension with native audio-video sync.The core scripts require muapi-cli. Install it once:
# via npm (recommended — no Python required)
npm install -g muapi-cli
# via pip
pip install muapi-cli
# or run without installing
npx muapi-cli --help
# Interactive setup
muapi auth configure
# Or pass directly
muapi auth configure --api-key "YOUR_MUAPI_KEY"
# Get your key at https://muapi.ai/dashboard
# Install all skills to your AI agent
npx skills add SamurAIGPT/Generative-Media-Skills --all
# Or install a specific skill
npx skills add SamurAIGPT/Generative-Media-Skills --skill muapi-media-generation
# Install to specific agents
npx skills add SamurAIGPT/Generative-Media-Skills --all -a claude-code -a cursor
muapi image generate "a cyberpunk city at night" --model flux-dev
# Download the result automatically
muapi image generate "a sunset over mountains" --model hidream-fast --download ./outputs
# Extract just the URL (agent-friendly)
muapi image generate "product on white bg" --model flux-schnell --output-json --jq '.outputs[0]'
# Use Nano-Banana reasoning to generate a 2K masterpiece
bash library/visual/nano-banana/scripts/generate-nano-art.sh \
--file ./my-source-image.jpg \
--subject "a glass hummingbird" \
--style "macro photography" \
--resolution "2k" \
--view
cd library/motion/cinema-director
# Create a 10-second epic reveal
bash scripts/generate-film.sh \
--subject "a cybernetic dragon over Tokyo" \
--intent "epic" \
--model "kling-v3.0-pro" \
--duration 10 \
--view
# Animate a reference image into video
bash library/motion/seedance-2/scripts/generate-seedance.sh \
--mode i2v \
--file ./concept.jpg \
--subject "camera slowly pulls back to reveal the full landscape" \
--intent "reveal" \
--view
# Extend an existing video
bash library/motion/seedance-2/scripts/generate-seedance.sh \
--mode extend \
--request-id "YOUR_REQUEST_ID" \
--subject "camera continues pulling back to reveal the vast city" \
--duration 10
Run muapi as a Model Context Protocol server so Claude Desktop, Cursor, or any MCP-compatible agent can call generation tools directly — no shell scripts needed.
muapi mcp serve
Claude Desktop config (~/Library/Application Support/Claude/claude_desktop_config.json):
{
"mcpServers": {
"muapi": {
"command": "muapi",
"args": ["mcp", "serve"],
"env": { "MUAPI_API_KEY": "your-key-here" }
}
}
}
This exposes 19 structured tools with full JSON Schema input/output definitions:
| Tool | Description |
|------|-------------|
| muapi_image_generate | Text-to-image (14 models) |
| muapi_image_edit | Image-to-image editing (11 models) |
| muapi_video_generate | Text-to-video (13 models) |
| muapi_video_from_image | Image-to-video (16 models) |
| muapi_audio_create | Music generation (Suno) |
| muapi_audio_from_text | Sound effects (MMAudio) |
| muapi_enhance_upscale | AI upscaling |
| muapi_enhance_bg_remove | Background removal |
| muapi_enhance_face_swap | Face swap image/video |
| muapi_enhance_ghibli | Ghibli style transfer |
| muapi_edit_lipsync | Lip sync to audio |
| muapi_edit_clipping | AI highlight extraction |
| muapi_predict_result | Poll prediction status |
| muapi_upload_file | Upload local file → URL |
| muapi_keys_list | List API keys |
| muapi_keys_create | Create API key |
| muapi_keys_delete | Delete API key |
| muapi_account_balance | Get credit balance |
| muapi_account_topup | Add credits (Stripe checkout) |
# Submit async, capture request_id, poll when ready
REQUEST_ID=$(muapi video generate "a dog running on a beach" \
--model kling-master --no-wait --output-json --jq '.request_id' | tr -d '"')
# ... do other work ...
muapi predict wait "$REQUEST_ID" --download ./outputs
# Pipe a prompt from another command
generate_prompt | muapi image generate - --model flux-dev
# Chain: upload → edit → download
URL=$(muapi upload file ./photo.jpg --output-json --jq '.url' | tr -d '"')
muapi image edit "make it look like a painting" --image "$URL" \
--model flux-kontext-pro --download ./outputs
This repository includes a streamlined schema_data.json that core scripts use at runtime to:
aspect_ratio, resolution, and duration values.Discover all available models via the CLI:
muapi models list
muapi models list --category video --output-json
Optimized for the next generation of AI development environments:
--output-json, --jq, semantic exit codes for scripting.MIT © 2026