@krishna-505

SkillCompass OC Canary (Internal)

SkillCompass plugin for OpenClaw — skill quality evaluation, usage tracking, and inbox suggestions

当前版本
v1.1.0-oc.8
code-plugin社区source-linked
<h1 align="center">SkillCompass</h1> <p align="center"> <strong>Evaluate quality. Find the weakest link. Fix it. Prove it worked. Repeat.</strong> </p> <p align="center"> <a href="https://github.com/Evol-ai/SkillCompass">GitHub</a> &middot; <a href="SKILL.md">SKILL.md</a> &middot; <a href="schemas/">Schemas</a> &middot; <a href="CHANGELOG.md">Changelog</a> </p> <p align="center"> <a href="https://clawhub.ai/skill/skill-compass"><img src="https://img.shields.io/badge/ClawHub-skill--compass-orange.svg" alt="ClawHub" /></a> <img src="https://img.shields.io/badge/license-MIT-blue.svg" alt="MIT License" /> <img src="https://img.shields.io/badge/node-%3E%3D18-brightgreen.svg" alt="Node >= 18" /> <img src="https://img.shields.io/badge/model-Claude%20Opus%204.6-purple.svg" alt="Claude Opus 4.6" /> </p>
What it isA local-first skill quality evaluator and management tool for Claude Code / OpenClaw. Six-dimension scoring, usage-driven suggestions, guided improvement, version tracking.
Pain it solvesTurns "tweak and hope" into diagnose → targeted fix → verified improvement. Turns "install and forget" into ongoing visibility over what's working, what's stale, and what's risky.
Use in 30 seconds/skillcompass — see your skill health at a glance. /eval-skill {path} — instant quality report showing exactly what's weakest and what to improve next.

Evaluate → find weakest link → fix it → prove it worked → next weakness → repeat. Meanwhile, Skill Inbox watches your usage and tells you what needs attention.


Who This Is For

<table> <tr><td width="50%">

For

  • Anyone maintaining agent skills and wanting measurable quality
  • Developers who want directed improvement — not guesswork, but knowing exactly which dimension to fix next
  • Teams needing a quality gate — any tool that edits a skill gets auto-evaluated
  • Users who install many skills and need visibility over what's actually used, what's stale, and what's risky
</td><td>

Not For

  • General code review or runtime debugging
  • Creating new skills from scratch (use skill-creator)
  • Evaluating non-skill files
</td></tr> </table>

Quick Start

Prerequisites: Claude Opus 4.6 (complex reasoning + consistent scoring) · Node.js v18+ (local validators)

Claude Code

git clone https://github.com/Evol-ai/SkillCompass.git
cd SkillCompass && npm install

# User-level (all projects)
rsync -a --exclude='.git'  . ~/.claude/skills/skill-compass/

# Or project-level (current project only)
rsync -a --exclude='.git'  . .claude/skills/skill-compass/

First run: SkillCompass auto-triggers a brief onboarding — scans your installed skills (~5 seconds), offers statusLine setup, then hands control back. Claude Code will request permission for node commands; select "Allow always" to avoid repeated prompts.

OpenClaw

git clone https://github.com/Evol-ai/SkillCompass.git
cd SkillCompass && npm install
# Follow OpenClaw skill installation docs for your setup
rsync -a --exclude='.git'  . <your-openclaw-skills-path>/skill-compass/

If your OpenClaw skills live outside the default scan roots, add them to skills.load.extraDirs in ~/.openclaw/openclaw.json:

{
  "skills": {
    "load": {
      "extraDirs": ["<your-openclaw-skills-path>"]
    }
  }
}

Usage

/skillcompass is the single entry point. Use it with a slash command or just talk naturally — both work:

/skillcompass                              → see what needs attention
/skillcompass evaluate my-skill            → six-dimension quality report
"improve the nano-banana skill"            → fix weakest dimension, verify, next
"what skills haven't I used recently?"     → usage-based insights
"security scan this skill"                 → D3 security deep-dive

What It Does

<p align="center"> <img src="assets/skill-quality-report.png" alt="SkillCompass — Skill Quality Report" width="380" /> </p>

The score isn't the point — the direction is. You instantly see which dimension is the bottleneck and what to do about it.

Each /eval-improve round follows a closed loop: fix the weakest → re-evaluate → verify improvement → next weakest. No fix is saved unless the re-evaluation confirms it actually helped.


Six-Dimension Evaluation Model

IDDimensionWeightWhat it evaluates
D1Structure10%Frontmatter validity, markdown format, declarations
D2Trigger15%Activation quality, rejection accuracy, discoverability
D3Security20%Secrets, injection, permissions, exfiltration, embedded shell
D4Functional30%Core quality, edge cases, output stability, error handling
D5Comparative15%Value over direct prompting (with vs without skill)
D6Uniqueness10%Overlap with similar skills, model supersession risk
overall_score = round((D1×0.10 + D2×0.15 + D3×0.20 + D4×0.30 + D5×0.15 + D6×0.10) × 10)
VerdictCondition
PASSscore >= 70 AND D3 pass
CAUTION50–69, or D3 High findings
FAILscore < 50, or D3 Critical (gate override)

Skill Inbox — Usage-Driven Suggestions

SkillCompass passively tracks which skills you actually use and surfaces suggestions when something needs attention — unused skills, stale evaluations, declining usage, available updates, and more. 9 built-in rules, all based on real invocation data.

  • Suggestions have a lifecycle: pending → acted / snoozed / dismissed, with auto-reactivation when conditions change
  • All data stays local — no network calls unless you explicitly request updates
  • Tracking is automatic via hooks (~one line per skill invocation), zero configuration

Features

Evaluate → Improve → Verify

/eval-skill scores six dimensions and pinpoints the weakest. /eval-improve targets that dimension, applies a fix, and re-evaluates — only saves when the target dimension improved and security/functionality didn't regress. Then move to the next weakness.

Skill Lifecycle

SkillCompass covers the full lifecycle of your skills — not just one-time evaluation.

Install — auto-scans your inventory, quick-checks security patterns across packages and sub-skills.

Ongoing — usage hooks passively track every invocation. Skill Inbox turns this into actionable insights: which skills are never used, which are declining, which are heavily used but never evaluated, which have updates available.

On edit — hooks auto-check structure + security on every SKILL.md write through Claude. Catches injection, exfiltration, embedded shell. Warns, never blocks.

On change — SHA-256 snapshots ensure any version is recoverable. D3 or D4 regresses after improvement? Snapshot restored automatically.

On update — update checker reads local git state passively; network only when you ask. Three-way merge preserves your local improvements region-by-region.

Scale

One skill or fifty — same workflow. /eval-audit scans a whole directory and ranks results worst-first so you fix what matters most. /eval-evolve chains multiple improve rounds automatically (default 6, stops at PASS or plateau). --ci flag outputs machine-readable JSON with exit codes for pipeline integration.


Works With Everything

No point-to-point integration needed. The Pre-Accept Gate intercepts all SKILL.md edits regardless of source.

ToolHow it works togetherGuide
ClaudeceptionExtracts skill → auto-evaluation catches security holes + redundancy → directed fixguide
Self-Improving AgentLogs errors → feed as signals → SkillCompass maps to dimensions and fixesguide

Design Principles

  • Local-first: All data stays on your machine. No network calls except when you explicitly request updates.
  • Read-only by default: Evaluation and reporting are read-only. Write operations (improve, merge, rollback) require explicit opt-in.
  • Passive tracking, active decisions: Hooks collect usage data silently. Suggestions are surfaced, never auto-acted on.
  • Dual-channel UX: Keyboard-selectable choices for actions, natural language for queries. Both always available.

Feedback Signal Standard

SkillCompass defines an open feedback-signal.json schema for any tool to report skill usage data:

/eval-skill ./my-skill/SKILL.md --feedback ./feedback-signals.json

Signals: trigger_accuracy, correction_count, correction_patterns, adoption_rate, ignore_rate, usage_frequency. The schema is extensible (additionalProperties: true) — any pipeline can produce or consume this format.


License

MIT — Use, modify, distribute freely. See LICENSE for details.

源码与版本

源码仓库

Evol-ai/SkillCompass

打开仓库

源码提交

230fcdfe28570c5fa258658ba47be8024c53cd17

查看提交

安装命令

openclaw plugins install clawhub:skillcompass-oc-canary

元数据

  • 包名: skillcompass-oc-canary
  • 创建时间: 2026/04/09
  • 更新时间: 2026/04/10
  • 执行代码:
  • 源码标签: feature/openclaw

兼容性

  • 构建于 OpenClaw: 2026.3.24-beta.2
  • 插件 API 范围: >=2026.3.24-beta.2
  • 标签: canary, latest
  • 文件数: 56