✓ Verified 💻 Development ✓ Enhanced Data

Token Saver 75plus

Always-on token optimization + model routing protocol.

Rating
5 (419 reviews)
Downloads
1,014 downloads
Version
1.0.0

Overview

Always-on token optimization + model routing protocol.

Complete Documentation

View Source →

Token Saver 75+ with Model Routing

Core Principle

Understand fully, execute cheaply. The orchestrator must fully understand the task before routing. Never sacrifice comprehension for speed.

Request Classifier (silent, every message)

TierPatternOrchestratorExecutor
T1yes/no, status, trivial facts, quick lookupsHandle alone
T2summaries, how-to, lists, bulk processing, formattingHandle alone OR spawn GroqGroq (FREE)
T3debugging, multi-step, code generation, structured analysisOrchestrate + spawnCodex for code, Groq for bulk
T4strategy, complex decisions, multi-agent coordination, creativeSpawn OpusOpus orchestrates, spawns Codex/Groq from within

Model Routing Table

ModelUse ForCostSpawn with
groq/llama-3.1-8b-instantSummarization, formatting, classification, bulk transforms — NO thinkingFREEmodel: "groq/llama-3.1-8b-instant"
openai/gpt-5.3-codexALL code generation, code review, refactoring$$$model: "openai/gpt-5.3-codex"
openai/gpt-5.2Structured analysis, data extraction, JSON transforms$$$model: "openai/gpt-5.2"
anthropic/claude-opus-4-6Strategy, complex orchestration, failure recovery (T4 only)$$$$model: "anthropic/claude-opus-4-6"

Routing via sessions_spawn

When to spawn (MANDATORY)

  • Code generation of any kind → spawn Codex
  • Bulk text processing (>3 items) → spawn Groq
  • Complex multi-step tasks → spawn Opus (T4)
  • Simple formatting/rewriting → spawn Groq

When NOT to spawn

  • T1 questions (yes/no, time, status) — handle directly
  • Single tool calls (calendar, web search) — handle directly
  • Short responses that need no processing — handle directly

Spawn patterns

Groq (free bulk work):

text
sessions_spawn(
  task: "<clear instruction with all context included>",
  model: "groq/llama-3.1-8b-instant"
)

Codex (all code):

text
sessions_spawn(
  task: "Write <language> code that <detailed spec>. Include comments. Output the complete file.",
  model: "openai/gpt-5.3-codex"
)

Opus (T4 strategy):

text
sessions_spawn(
  task: "<full context + goal>. You have full tool access. Use sessions_spawn with Codex for code and Groq for bulk subtasks.",
  model: "anthropic/claude-opus-4-6"
)

Critical spawn rules

  • Include ALL context in the task string — spawned agents have no conversation history
  • Be specific — vague tasks waste tokens on clarification
  • One task per spawn — don't bundle unrelated work
  • For code: always use Codex — never write code yourself

Output Compression (applies to ALL tiers, ALL models)

Templates

  • STATUS: OK/WARN/FAIL one-liner
  • CHOICE: A vs B → Recommend: X (1 line why)
  • CAUSE→FIX→VERIFY: 3 bullets max
  • RESULT: data/output directly, no wrap-up

Rules

  • No filler. No restating the question. Lead with the answer.
  • Bullets/tables/code > prose.
  • Do not narrate routine tool calls.
  • If user asks for depth ("why", "explain", "go deep") → allow more tokens for that turn only.

Budget by tier

TierMax output
T11-3 lines
T25-15 bullets
T3Structured sections, <400 words
T4Longer allowed, still dense

Tool Gating (before ANY tool call)

  • Already known? → No tool.
  • Batchable? → Parallelize.
  • Can a spawned Groq handle it? → Spawn instead of doing it yourself.
  • Cheapest path? → memory_search > partial read > full read > web.
  • Needed? → Do not fetch "just in case."

Failure Protocol

  • If Groq spawn fails → retry with GPT-5.2
  • If Codex spawn fails → retry with GPT-5.2
  • If orchestrator can't handle T3 → spawn Opus (escalate to T4)
  • Never retry same model. Escalate.

Measurement (when asked or during testing)

Append: [~X tokens | Tier: Tn | Route: model(s) used]

Installation

Terminal bash

openclaw install token-saver-75plus
    
Copied!

💻Code Examples

**Groq (free bulk work):**

groq-free-bulk-work.txt
sessions_spawn(
  task: "<clear instruction with all context included>",
  model: "groq/llama-3.1-8b-instant"
)

**Codex (all code):**

codex-all-code.txt
sessions_spawn(
  task: "Write <language> code that <detailed spec>. Include comments. Output the complete file.",
  model: "openai/gpt-5.3-codex"
)

**Opus (T4 strategy):**

opus-t4-strategy.txt
sessions_spawn(
  task: "<full context + goal>. You have full tool access. Use sessions_spawn with Codex for code and Groq for bulk subtasks.",
  model: "anthropic/claude-opus-4-6"
)

Tags

#web_and-frontend-development

Quick Info

Category Development
Model Claude 3.5
Complexity One-Click
Author mariovallereyes
Last Updated 3/10/2026
🚀
Optimized for
Claude 3.5
🧠

Ready to Install?

Get started with this skill in seconds

openclaw install token-saver-75plus