mulerouter
Generates images and videos using MuleRouter or MuleRun multimodal APIs. Text-to-Image, Image-to-Image, Text-to-Video, Image-to-Video, video editing (VACE, keyframe interpolation). Use when the user wants to generate, edit, or transform images and videos using AI models like Wan2.6 or Nano Banana.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/mulerouter/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How mulerouter Compares
| Feature / Agent | mulerouter | Standard Approach |
|---|---|---|
| Platform Support | multi | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Generates images and videos using MuleRouter or MuleRun multimodal APIs. Text-to-Image, Image-to-Image, Text-to-Video, Image-to-Video, video editing (VACE, keyframe interpolation). Use when the user wants to generate, edit, or transform images and videos using AI models like Wan2.6 or Nano Banana.
Which AI agents support this skill?
This skill is compatible with multi.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# MuleRouter API
Generate images and videos using MuleRouter or MuleRun multimodal APIs.
## Configuration Check
Before running any commands, verify the environment is configured:
### Step 1: Check for existing configuration
```bash
# Check environment variables
echo "MULEROUTER_SITE: $MULEROUTER_SITE"
echo "MULEROUTER_API_KEY: ${MULEROUTER_API_KEY:+[SET]}"
# Check for .env file
ls -la .env 2>/dev/null || echo "No .env file found"
```
### Step 2: Configure if needed
**Option A: Environment variables**
```bash
export MULEROUTER_SITE="mulerun" # or "mulerouter"
export MULEROUTER_API_KEY="your-api-key"
```
**Option B: Create .env file**
Create `.env` in the current working directory:
```env
MULEROUTER_SITE=mulerun
MULEROUTER_API_KEY=your-api-key
```
**Note:** The tool only reads `.env` from the current directory. Run scripts from the skill root (`skills/mulerouter-skills/`).
### Step 3: Using `uv` to run scripts
The skill uses `uv` for dependency management and execution. Make sure `uv` is installed and available in your PATH.
Run `uv sync` to install dependencies.
## Quick Start
### 1. List available models
```bash
uv run python scripts/list_models.py
```
### 2. Check model parameters
```bash
uv run python models/alibaba/wan2.6-t2v/generation.py --list-params
```
### 3. Generate content
**Text-to-Video:**
```bash
uv run python models/alibaba/wan2.6-t2v/generation.py --prompt "A cat walking through a garden"
```
**Text-to-Image:**
```bash
uv run python models/alibaba/wan2.6-t2i/generation.py --prompt "A serene mountain lake"
```
**Image-to-Video:**
```bash
uv run python models/alibaba/wan2.6-i2v/generation.py --prompt "Gentle zoom in" --images '["https://example.com/photo.jpg"]'
```
## Workflow
1. Check configuration: verify `MULEROUTER_SITE` and `MULEROUTER_API_KEY` are set
2. Install dependencies: run `uv sync`
3. Run `uv run python scripts/list_models.py` to discover available models
4. Run `uv run python models/<path>/<action>.py --list-params` to see parameters
5. Execute with appropriate parameters
6. Parse output URLs from results
## Tips
1. For an image generation model, a suggested timeout is 5 minutes.
2. For a video generation model, a suggested timeout is 15 minutes.
## References
- [REFERENCE.md](references/REFERENCE.md) - API configuration and CLI options
- [MODELS.md](references/MODELS.md) - Complete model specifications