✓ Verified
💻 Development
✓ Enhanced Data
Social Trust Manipulation Detector
Helps identify coordinated social trust manipulation in agent marketplaces — catching reputation gam
- Rating
- 4 (488 reviews)
- Downloads
- 689 downloads
- Version
- 1.0.0
Overview
Helps identify coordinated social trust manipulation in agent marketplaces — catching reputation gaming.
Complete Documentation
View Source →name: social-trust-manipulation-detector description: > Helps identify coordinated social trust manipulation in agent marketplaces — catching reputation gaming through sockpuppet networks, coordinated upvoting, and manufactured community signals that make unsafe skills appear trusted. version: 1.0.0 metadata: openclaw: requires: bins: [curl, python3] env: [] emoji: "🎭" agent_card: capabilities: [social-trust-manipulation-detection, sockpuppet-network-detection, coordinated-upvoting-detection, reputation-gaming-analysis] attack_surface: [L1] trust_dimension: rule-adoption published: clawhub: false moltbook: false
Your Trust Score Is Real. The Signals Behind It Are Manufactured.
Helps identify when a skill's trust reputation is built on coordinated
social manipulation rather than genuine community validation.
Problem
Trust in agent marketplaces flows through social signals: upvotes, downloads, comments, and follow counts. These signals are valuable precisely because they aggregate distributed judgment — when thousands of independent users find a skill useful and safe, their collective assessment carries real information. The assumption of independence is the attack surface. A coordinated network of accounts can manufacture the appearance of distributed consensus. A skill with 500 upvotes from a bot network looks identical to a skill with 500 upvotes from 500 independent developers. The marketplace's reputation system cannot distinguish manufactured trust from earned trust — and neither can most agents that rely on reputation as a trust signal. Social trust manipulation is the third pillar of the trust attack surface, alongside technical attacks (code injection) and structural attacks (supply chain compromise). It is the most scalable: a well-constructed sockpuppet network can manufacture trust faster than any code-level auditing can catch it, and the manufactured trust persists long after the network is dismantled. Legitimate skills earn trust gradually, from a diverse user base, with engagement patterns that correlate with actual skill utility. Manipulated skills earn trust in coordinated bursts, from accounts with suspicious creation patterns, with engagement that does not correlate with usage or outcomes.What This Detects
This detector examines social trust integrity across five dimensions:- Engagement velocity anomalies — Does the skill's vote/download
- Account cohort analysis — Do the skill's early upvoters share
- Engagement-to-utility correlation — Do social signals correlate with
- Cross-publisher coordination — Do multiple publishers in a marketplace
- Review authenticity signals — Do comments and reviews on the skill
How to Use
Input: Provide one of:- A skill identifier to assess the authenticity of its trust signals
- A publisher account to analyze for coordinated network membership
- A set of skills to assess for cross-publisher coordination patterns
- Engagement velocity analysis (organic vs. burst pattern)
- Account cohort fingerprint assessment
- Engagement-to-utility correlation score
- Cross-publisher coordination indicators
- Review authenticity assessment
- Manipulation verdict: AUTHENTIC / SUSPICIOUS / COORDINATED / MANUFACTURED
Example
Input: Assess social trust integrity forai-assistant-toolkit publisher
``
🎭 SOCIAL TRUST MANIPULATION ASSESSMENT
Publisher: ai-assistant-toolkit
Skills assessed: 4 (productivity-suite, auto-responder, data-fetcher, doc-reader)
Audit timestamp: 2025-09-05T12:00:00Z
Engagement velocity:
productivity-suite: 0 → 847 upvotes in 72 hours of launch ⚠️
auto-responder: 0 → 623 upvotes in 48 hours of launch ⚠️
data-fetcher: 0 → 412 upvotes in 60 hours of launch ⚠️
Organic baseline for comparable skills: 15-40 upvotes in first 72h
→ Burst pattern detected across all 4 skills
Account cohort analysis:
First 200 upvoters on productivity-suite:
Accounts created within 30-day window: 156/200 (78%) ⚠️
Cross-voting with auto-responder upvoters: 143/200 (71.5%) ⚠️
Accounts with no other skill interactions: 168/200 (84%) ⚠️
→ Sockpuppet cohort fingerprint detected
Engagement-to-utility correlation:
productivity-suite: 847 upvotes, 23 installs (ratio: 36.8:1) ⚠️
auto-responder: 623 upvotes, 18 installs (ratio: 34.6:1) ⚠️
Organic baseline ratio: 2:1 to 8:1 for comparable skills
→ Upvote-to-install ratio 4-18x above organic baseline
Cross-publisher coordination:
ai-assistant-toolkit upvoter network also upvoted:
fastcoder-pro (different publisher): 89% overlap ⚠️
quick-deploy-kit (different publisher): 76% overlap ⚠️
→ Mutual support network detected across 3 publishers
Review authenticity:
Top 20 reviews analyzed:
Unique vocabulary: 34 terms (low for 20 reviews) ⚠️
Specificity: Generic praise, no feature-specific feedback
Phrasing patterns: "absolutely essential", "game-changer" × 7 reviews
Manipulation verdict: MANUFACTURED
All four skills show coordinated burst voting, sockpuppet cohort fingerprints,
upvote-to-install ratios far above organic baseline, and cross-publisher
mutual support network membership. Trust signals for this publisher's skills
do not represent independent community validation.
Recommended actions:
- Treat trust score as unauthenticated pending platform investigation
- Evaluate skills on technical merit only, disregarding social signals
- Report coordination pattern to marketplace moderators
- Flag fastcoder-pro and quick-deploy-kit for same network membership
- Apply technical audit (supply-chain, permission-creep) before any install
``
Related Tools
- clone-farm-detector — Detects content-level cloning for reputation gaming;
- publisher-identity-verifier — Verifies publisher identity integrity;
- trust-velocity-calculator — Quantifies trust decay from update velocity;
- blast-radius-estimator — Estimates propagation impact if a skill is
Limitations
Social trust manipulation detection depends on access to engagement metadata (account creation dates, cross-voting patterns, install counts) that many marketplaces do not expose through public APIs. Where metadata is limited, only velocity analysis and review text assessment are available, which reduces detection confidence. Burst voting patterns can result from legitimate causes: coordinated community launches, press coverage, or featured placement can all produce rapid engagement that resembles manufactured trust. The account cohort analysis relies on observable fingerprints and will miss well-resourced adversaries who age accounts and vary patterns. This tool identifies social trust signals that warrant investigation — it does not confirm manipulation, which requires access to platform-level data that only marketplace operators can verify.Installation
Terminal bash
openclaw install social-trust-manipulation-detector
Copied!
Tags
#coding_agents-and-ides
Quick Info
Category Development
Model Claude 3.5
Complexity Multi-Agent
Author andyxinweiminicloud
Last Updated 3/10/2026
🚀
Optimized for
Claude 3.5
Ready to Install?
Get started with this skill in seconds
openclaw install social-trust-manipulation-detector
Related Skills
✓ Verified
💻 Development
4claw
4claw — a moderated imageboard for AI agents.
🧠 Claude-Ready
)}
★ 4.4 (118)
↓ 4,990
v1.0.0
✓ Verified
💻 Development
Aap Passport
Agent Attestation Protocol - The Reverse Turing Test.
🧠 Claude-Ready
)}
★ 4.3 (89)
↓ 4,621
v1.0.0
✓ Verified
💻 Development
Acestep Lyrics Transcription
Transcribe audio to timestamped lyrics using OpenAI Whisper or ElevenLabs Scribe API.
⚡ GPT-Optimized
)}
★ 3.8 (274)
↓ 17,648
v1.0.0
✓ Verified
💻 Development
Adaptive Suite
A continuously adaptive skill suite that empowers Clawdbot.
🧠 Claude-Ready
)}
★ 4.7 (88)
↓ 1,625
v1.0.0