Zettel Link
This skill maintains the Note Embeddings for Zettelkasten, to search notes, retrieve notes, and disc
- Rating
- 4.4 (33 reviews)
- Downloads
- 982 downloads
- Version
- 1.0.0
Overview
This skill maintains the Note Embeddings for Zettelkasten, to search notes, retrieve notes, and discover.
✨Key Features
— Setup and Config
— Create Embeddings
— Semantic Search
— Semantic Connection Discovery
Complete Documentation
View Source →
Zettel Link Skill
This skill provides a suite of idempotent Python scripts to embed, search, and link notes in an Obsidian vault using semantic similarity. All scripts live in scripts/ and support multiple embedding providers.
The skill should be triggered when the user wants to search notes, retrieve notes, or discover connections between notes.
If the search directory is indexed with embeddings, the skill should prompt the user if they want to create new embeddings.
Dependencies
- uv 0.10.0+
- Python 3.10+
- One of the following embedding providers:
- Ollama with
mxbai-embed-large(local, default) - OpenAI API with
text-embedding-3-small - Google Gemini API with
text-embedding-004
Overview of Commands
uv run scripts/config.py: Configure the embedding model and other settings.uv run scripts/embed.py: Embed notes and cache to.embeddings/embeddings.jsonuv run scripts/search.py: Semantic search over embedded notesuv run scripts/link.py: Discover semantic connections, output to.embeddings/links.json
Workflow
Step 0 — Setup and Config
If the config/config.json file does not exist, create it:
uv run scripts/config.py
This creates config/config.json with defaults:
{
"model": "mxbai-embed-large",
"provider": {
"name": "ollama",
"url": "http://localhost:11434"
},
"max_input_length": 8192,
"cache_dir": ".embeddings",
"default_threshold": 0.65,
"top_k": 5,
"skip_dirs": [".obsidian", ".trash", ".embeddings", "Spaces", "templates"],
"skip_files": ["CLAUDE.md", "Vault.md", "Dashboard.md", "templates.md"]
}
To use a remote provider:
# OpenAI
uv run scripts/config.py --provider openai
# Gemini
uv run scripts/config.py --provider gemini
# Custom model
uv run scripts/config.py --provider openai --model text-embedding-3-large
To adjust tuning parameters:
uv run scripts/config.py --top-k 10 --threshold 0.7 --max-input-length 4096
Step 1 — Create Embeddings
uv run scripts/embed.py --input <directory>
This creates with the embedding cache.
- Incremental updates: Only re-embeds files that have been modified since the last run (based on file modification time).
- Text truncation: Automatically truncates text to
max_input_lengthbefore embedding. - Stale pruning: Removes entries for files that no longer exist.
- Force re-embed: Use
--forceto re-embed everything.
Step 2 — Semantic Search
uv run scripts/search.py --input <directory> --query "<query>"
This embeds the query using the configured provider and compares it with all cached embeddings, returning the top_k most similar notes.
Results are saved to .
Step 3 — Semantic Connection Discovery
uv run scripts/link.py --input <directory>
This computes cosine similarity for all note pairs and outputs connections above the default_threshold to .
The output includes:
- A flat list of all link pairs with scores
- A per-note grouping for easy lookup
--threshold to widen or narrow the connection discovery.Cache
- Format: JSON with metadata envelope (
metadata+data) - Location:
/.embeddings/embeddings.json - Metadata: Tracks generation timestamp, model, provider, embedding size
- Invalidation: Based on file modification time (
mtime) - Force rebuild: Delete the cache file or use
--forceflag
Agent Instructions
When using this skill:
- Always run
config.pyfirst ifconfig/config.jsondoes not exist. - Run
embed.pybeforesearch.pyorlink.py— the cache must exist. - For remote providers (openai, gemini), ensure the API key environment variable is set (or provide a local
.envfile in the skill directory). - All scripts are idempotent and safe to re-run.
Installation
openclaw install zettel-link
💻Code Examples
uv run scripts/embed.py --input <directory>
This creates `<directory>/.embeddings/embeddings.json` with the embedding cache.
- **Incremental updates**: Only re-embeds files that have been modified since the last run (based on file modification time).
- **Text truncation**: Automatically truncates text to `max_input_length` before embedding.
- **Stale pruning**: Removes entries for files that no longer exist.
- **Force re-embed**: Use `--force` to re-embed everything.
### Step 2 — Semantic Searchuv run scripts/search.py --input <directory> --query "<query>"
This embeds the query using the configured provider and compares it with all cached embeddings, returning the `top_k` most similar notes.
Results are saved to `<directory>/.embeddings/search_results.json`.
### Step 3 — Semantic Connection Discovery{
"model": "mxbai-embed-large",
"provider": {
"name": "ollama",
"url": "http://localhost:11434"
},
"max_input_length": 8192,
"cache_dir": ".embeddings",
"default_threshold": 0.65,
"top_k": 5,
"skip_dirs": [".obsidian", ".trash", ".embeddings", "Spaces", "templates"],
"skip_files": ["CLAUDE.md", "Vault.md", "Dashboard.md", "templates.md"]
}# OpenAI
uv run scripts/config.py --provider openai
# Gemini
uv run scripts/config.py --provider gemini
# Custom model
uv run scripts/config.py --provider openai --model text-embedding-3-largeTags
Quick Info
Ready to Install?
Get started with this skill in seconds
Related Skills
4claw
4claw — a moderated imageboard for AI agents.
Aap Passport
Agent Attestation Protocol - The Reverse Turing Test.
Adaptive Suite
A continuously adaptive skill suite that empowers Clawdbot.
Adversarial Prompting
Adversarial analysis to critique, fix.