✓ Verified
💻 Development
✓ Enhanced Data
Api Benchmark
Benchmark token generation speed across multiple LLM API providers.
- Rating
- 4 (356 reviews)
- Downloads
- 785 downloads
- Version
- 1.0.0
Overview
Benchmark token generation speed across multiple LLM API providers.
Complete Documentation
View Source →
API Token Speed Benchmark
This skill benchmarks token generation speed across multiple LLM API providers.
When to use this skill
Use this skill when you need to:
- Compare token generation speed across different API providers
- Measure latency and throughput of LLM models
- Verify API connectivity and authentication
- Test new API endpoints or models
How to run benchmarks
List available targets
bash
python3 main.py --targets
Run benchmark on a specific target
bash
python3 main.py run --label <target-label>
Run benchmark on all targets
bash
python3 main.py run --all
Run preflight check (verify API connectivity)
bash
python3 main.py check --label <target-label>
python3 main.py check --all
Options
-l, --label: Specific target label to benchmark-a, --all: Run on all available targets-r, --repeat: Number of runs per prompt level (default: 1)-c, --category: Run specific prompt category (can repeat: -c short -c medium). Options: short, medium, long-q, --quiet: Quiet mode - suppress progress output--timeout N: Request timeout in seconds (default: 120)--table: Output as formatted table (default: JSON)
Configuration
The tool reads configuration from ~/.openclaw/openclaw.json. Targets are defined in the models.providers section with:
baseUrl: API base URLapiKey: Authentication key (or${ENV_VAR}to read from environment variable)api: API format (anthropic-messages, openai-completions, openai-responses)models: List of model configurations
"apiKey": "${ANTHROPIC_API_KEY}"will read from theANTHROPIC_API_KEYenvironment variable
json
{
"models": {
"providers": {
"my-provider": {
"baseUrl": "https://api.example.com",
"apiKey": "sk-xxx",
"api": "openai-completions",
"models": [
{ "id": "model-name", "api": "openai-completions" }
]
}
}
}
}
Output Metrics
- TTFT (Time To First Token): Latency before first token arrives (seconds)
- TPS (Tokens Per Second): Generation throughput
- Total Time: Full generation duration (seconds)
- Input/Output Tokens: Token counts from API usage data (or estimated at 4 chars/token if not provided by API)
Example Usage
bash
# Check if a specific target is reachable
python3 main.py check --label my-provider
# Benchmark a single target
python3 main.py run --label my-provider --repeat 3
# Compare all targets
python3 main.py run --all --table
Installation
Terminal bash
openclaw install api-benchmark
Copied!
💻Code Examples
python3 main.py check --all
python3-mainpy-check---all.txt
### Options
- `-l, --label`: Specific target label to benchmark
- `-a, --all`: Run on all available targets
- `-r, --repeat`: Number of runs per prompt level (default: 1)
- `-c, --category`: Run specific prompt category (can repeat: -c short -c medium). Options: short, medium, long
- `-q, --quiet`: Quiet mode - suppress progress output
- `--timeout N`: Request timeout in seconds (default: 120)
- `--table`: Output as formatted table (default: JSON)
## Configuration
The tool reads configuration from `~/.openclaw/openclaw.json`. Targets are defined in the `models.providers` section with:
- `baseUrl`: API base URL
- `apiKey`: Authentication key (or `${ENV_VAR}` to read from environment variable)
- `api`: API format (anthropic-messages, openai-completions, openai-responses)
- `models`: List of model configurations
**Security Note**: Instead of hardcoding API keys in the config file, use environment variable placeholders:
- `"apiKey": "${ANTHROPIC_API_KEY}"` will read from the `ANTHROPIC_API_KEY` environment variable
Example provider config:}
.txt
## Output Metrics
- **TTFT** (Time To First Token): Latency before first token arrives (seconds)
- **TPS** (Tokens Per Second): Generation throughput
- **Total Time**: Full generation duration (seconds)
- **Input/Output Tokens**: Token counts from API usage data (or estimated at 4 chars/token if not provided by API)
Note: Token counts are reported by the API when available. If the API doesn't return token counts, they are estimated at 4 characters per token.
## Example Usageexample.json
{
"models": {
"providers": {
"my-provider": {
"baseUrl": "https://api.example.com",
"apiKey": "sk-xxx",
"api": "openai-completions",
"models": [
{ "id": "model-name", "api": "openai-completions" }
]
}
}
}
}example.sh
# Check if a specific target is reachable
python3 main.py check --label my-provider
# Benchmark a single target
python3 main.py run --label my-provider --repeat 3
# Compare all targets
python3 main.py run --all --table⚙️Configuration Options
| Option | Type | Default | Description |
|---|---|---|---|
ANTHROPIC_API_KEY | string | - | environment variable |
Tags
#coding_agents-and-ides
#api
Quick Info
Category Development
Model Claude 3.5
Complexity One-Click
Author polarjunction
Last Updated 3/10/2026
🚀
Optimized for
Claude 3.5
Ready to Install?
Get started with this skill in seconds
openclaw install api-benchmark
Related Skills
✓ Verified
💻 Development
4claw
4claw — a moderated imageboard for AI agents.
🧠 Claude-Ready
)}
★ 4.4 (118)
↓ 4,990
v1.0.0
✓ Verified
💻 Development
Aap Passport
Agent Attestation Protocol - The Reverse Turing Test.
🧠 Claude-Ready
)}
★ 4.3 (89)
↓ 4,621
v1.0.0
✓ Verified
💻 Development
Acestep Lyrics Transcription
Transcribe audio to timestamped lyrics using OpenAI Whisper or ElevenLabs Scribe API.
⚡ GPT-Optimized
)}
★ 3.8 (274)
↓ 17,648
v1.0.0
✓ Verified
💻 Development
Adaptive Suite
A continuously adaptive skill suite that empowers Clawdbot.
🧠 Claude-Ready
)}
★ 4.7 (88)
↓ 1,625
v1.0.0