Mulerouter
Generates images and videos using MuleRouter or MuleRun multimodal APIs.
- Rating
- 4.1 (322 reviews)
- Downloads
- 34,387 downloads
- Version
- 1.0.0
Overview
Generates images and videos using MuleRouter or MuleRun multimodal APIs.
✨Key Features
Check for existing configuration
Configure if needed
Using `uv` to run scripts
Complete Documentation
View Source →
MuleRouter API
Generate images and videos using MuleRouter or MuleRun multimodal APIs.
Required Environment Variables
This skill requires the following environment variables to be set before use:
| Variable | Required | Description |
|---|---|---|
| MULEROUTER_API_KEY | Yes | API key for authentication (get one here) |
| MULEROUTER_BASE_URL | Yes | Custom API base URL (e.g., https://api.mulerouter.ai). Takes priority over SITE. |
| MULEROUTER_SITE | Yes | API site: mulerouter or mulerun. Used if BASE_URL is not set. |
MULEROUTER_BASE_URL or MULEROUTER_SITE must be set.The API key is included in Authorization: Bearer headers when making network calls to the configured API endpoint.
If any of these variables are missing, the scripts will fail with a configuration error. Check the Configuration section below to set them up.
Configuration Check
Before running any commands, verify the environment is configured:
Step 1: Check for existing configuration
Run the built-in config check script:
uv run python -c "from core.config import load_config; load_config(); print('Configuration OK')"
If this prints "Configuration OK", skip to Step 3. If it raises a ValueError, proceed to Step 2.
Step 2: Configure if needed
If the variables above are not set, ask the user to provide their API key and preferred endpoint.
Create a .env file in the skill's working directory:
# Option 1: Use custom base URL (takes priority over SITE)
MULEROUTER_BASE_URL=https://api.mulerouter.ai
MULEROUTER_API_KEY=your-api-key
# Option 2: Use site (if BASE_URL not set)
# MULEROUTER_SITE=mulerun
# MULEROUTER_API_KEY=your-api-key
Note: MULEROUTER_BASE_URL takes priority over MULEROUTER_SITE. If both are set, MULEROUTER_BASE_URL is used.
Note: The skill only loads variables prefixed with MULEROUTER_ from the .env file. Other variables in the file are ignored.
Important: Do NOT use export shell commands to set credentials. Use a .env file or ensure the variables are already present in your shell environment before invoking the skill.
Step 3: Using uv to run scripts
The skill uses uv for dependency management and execution. Make sure uv is installed and available in your PATH.
Run uv sync to install dependencies.
Quick Start
1. List available models
uv run python scripts/list_models.py
2. Check model parameters
uv run python models/alibaba/wan2.6-t2v/generation.py --list-params
3. Generate content
Text-to-Video:
uv run python models/alibaba/wan2.6-t2v/generation.py --prompt "A cat walking through a garden"
Text-to-Image:
uv run python models/alibaba/wan2.6-t2i/generation.py --prompt "A serene mountain lake"
Image-to-Video:
uv run python models/alibaba/wan2.6-i2v/generation.py --prompt "Gentle zoom in" --image "https://example.com/photo.jpg" #remote image url
uv run python models/alibaba/wan2.6-i2v/generation.py --prompt "Gentle zoom in" --image "/path/to/local/image.png" #local image path
Image Input
For image parameters (--image, --images, etc.), prefer local file paths over base64.
# Preferred: local file path (auto-converted to base64)
--image /tmp/photo.png
--images ["/tmp/photo.png"]
Local file paths are validated before reading: only files with recognized image extensions (.png, .jpg, .jpeg, .gif, .bmp, .webp, .tiff, .tif, .svg, .ico, .heic, .heif, .avif) are accepted. Paths pointing to sensitive system directories or non-image files are rejected. Valid image files are converted to base64 and sent to the API, avoiding command-line length limits that occur with raw base64 strings.
Workflow
- Check configuration: verify
MULEROUTER_API_KEYand eitherMULEROUTER_BASE_URLorMULEROUTER_SITEare set - Install dependencies: run
uv sync - Run
uv run python scripts/list_models.pyto discover available models - Run
uv run python models/to see parameters/ .py --list-params - Execute with appropriate parameters
- Parse output URLs from results
Model Selection
When listing models, each model's tags (e.g., [SOTA]) are displayed by default next to its name. Tags help identify model characteristics at a glance — for example, SOTA indicates a state-of-the-art model.
You can also filter models by tag using --tag:
uv run python scripts/list_models.py --tag SOTA
If you are unsure which model to use, present the available options to the user and let them choose. Use the AskUserQuestion tool (or equivalent interactive prompt) to ask the user which model they prefer. For example, if the user asks to "generate an image" without specifying a model, list the relevant image generation models with their tags and descriptions, and ask the user to pick one.
Tips
- For an image generation model, a suggested timeout is 5 minutes.
- For a video generation model, a suggested timeout is 15 minutes.
References
- REFERENCE.md - API configuration and CLI options
- MODELS.md - Complete model specifications
Installation
openclaw install mulerouter
💻Code Examples
uv run python -c "from core.config import load_config; load_config(); print('Configuration OK')"
If this prints "Configuration OK", skip to **Step 3**. If it raises a `ValueError`, proceed to Step 2.
### Step 2: Configure if needed
**If the variables above are not set**, ask the user to provide their API key and preferred endpoint.
**Create a `.env` file** in the skill's working directory:# MULEROUTER_API_KEY=your-api-key
**Note:** `MULEROUTER_BASE_URL` takes priority over `MULEROUTER_SITE`. If both are set, `MULEROUTER_BASE_URL` is used.
**Note:** The skill only loads variables prefixed with `MULEROUTER_` from the `.env` file. Other variables in the file are ignored.
**Important:** Do NOT use `export` shell commands to set credentials. Use a `.env` file or ensure the variables are already present in your shell environment before invoking the skill.
### Step 3: Using `uv` to run scripts
The skill uses `uv` for dependency management and execution. Make sure `uv` is installed and available in your PATH.
Run `uv sync` to install dependencies.
## Quick Start
### 1. List available modelsuv run python models/alibaba/wan2.6-t2v/generation.py --list-params
### 3. Generate content
**Text-to-Video:**uv run python models/alibaba/wan2.6-i2v/generation.py --prompt "Gentle zoom in" --image "/path/to/local/image.png" #local image path
## Image Input
For image parameters (`--image`, `--images`, etc.), **prefer local file paths** over base64.--images ["/tmp/photo.png"]
Local file paths are validated before reading: only files with recognized image extensions (`.png`, `.jpg`, `.jpeg`, `.gif`, `.bmp`, `.webp`, `.tiff`, `.tif`, `.svg`, `.ico`, `.heic`, `.heif`, `.avif`) are accepted. Paths pointing to sensitive system directories or non-image files are rejected. Valid image files are converted to base64 and sent to the API, avoiding command-line length limits that occur with raw base64 strings.
## Workflow
1. Check configuration: verify `MULEROUTER_API_KEY` and either `MULEROUTER_BASE_URL` or `MULEROUTER_SITE` are set
2. Install dependencies: run `uv sync`
3. Run `uv run python scripts/list_models.py` to discover available models
4. Run `uv run python models/<path>/<action>.py --list-params` to see parameters
5. Execute with appropriate parameters
6. Parse output URLs from results
## Model Selection
When listing models, each model's **tags** (e.g., `[SOTA]`) are displayed by default next to its name. Tags help identify model characteristics at a glance — for example, `SOTA` indicates a state-of-the-art model.
You can also filter models by tag using `--tag`:# Option 1: Use custom base URL (takes priority over SITE)
MULEROUTER_BASE_URL=https://api.mulerouter.ai
MULEROUTER_API_KEY=your-api-key
# Option 2: Use site (if BASE_URL not set)
# MULEROUTER_SITE=mulerun
# MULEROUTER_API_KEY=your-api-key# Preferred: local file path (auto-converted to base64)
--image /tmp/photo.png
--images ["/tmp/photo.png"]Tags
Quick Info
Ready to Install?
Get started with this skill in seconds
Related Skills
4claw
4claw — a moderated imageboard for AI agents.
Aap Passport
Agent Attestation Protocol - The Reverse Turing Test.
Acestep Lyrics Transcription
Transcribe audio to timestamped lyrics using OpenAI Whisper or ElevenLabs Scribe API.
Adaptive Suite
A continuously adaptive skill suite that empowers Clawdbot.