tessl-skill-review

Evaluate, score, and review an Agent Skill or SKILL.md using Tessl as the primary evaluator. Use when asked to measure skill quality, score a skill, review a skill against best practices, compare before/after skill revisions, or generate structured improvement feedback for a skill directory or SKILL.md file.

33 stars

Best use case

tessl-skill-review is best used when you need a repeatable AI agent workflow instead of a one-off prompt.

Evaluate, score, and review an Agent Skill or SKILL.md using Tessl as the primary evaluator. Use when asked to measure skill quality, score a skill, review a skill against best practices, compare before/after skill revisions, or generate structured improvement feedback for a skill directory or SKILL.md file.

Teams using tessl-skill-review should expect a more consistent output, faster repeated execution, less prompt rewriting.

When to use this skill

  • You want a reusable workflow that can be run more than once with consistent structure.

When not to use this skill

  • You only need a quick one-off answer and do not need a reusable workflow.
  • You cannot install or maintain the underlying files, dependencies, or repository context.

Installation

Claude Code / Cursor / Codex

$curl -o ~/.claude/skills/tessl-skill-review/SKILL.md --create-dirs "https://raw.githubusercontent.com/aAAaqwq/AGI-Super-Team/main/skills/tessl-skill-review/SKILL.md"

Manual Installation

  1. Download SKILL.md from GitHub
  2. Place it in .claude/skills/tessl-skill-review/SKILL.md inside your project
  3. Restart your AI agent — it will auto-discover the skill

How tessl-skill-review Compares

Feature / Agenttessl-skill-reviewStandard Approach
Platform SupportNot specifiedLimited / Varies
Context Awareness High Baseline
Installation ComplexityUnknownN/A

Frequently Asked Questions

What does this skill do?

Evaluate, score, and review an Agent Skill or SKILL.md using Tessl as the primary evaluator. Use when asked to measure skill quality, score a skill, review a skill against best practices, compare before/after skill revisions, or generate structured improvement feedback for a skill directory or SKILL.md file.

Where can I find the source code?

You can find the source code on GitHub using the link provided at the top of the page.

SKILL.md Source

# Tessl Skill Review

Use Tessl as the default scoring engine for skill quality review.

This skill is for **measuring quality**, not just giving vibes. Prefer Tessl-backed review first, then add your own judgment on top.

## Primary use cases

Use this skill when asked to:
- score a skill
- evaluate a `SKILL.md`
- review a skill against best practices
- compare two versions of a skill
- decide whether a skill is ready to publish
- find weaknesses in skill triggering, structure, or instructions

## Core workflow

### 1) Identify the review target

Accept either:
- a skill directory containing `SKILL.md`
- a direct path to `SKILL.md`
- a repo path with one or more skills to audit

If the request is ambiguous, clarify which skill or directory to score.

### 2) Prefer Tessl review first

If Tessl CLI is available, start with:

```bash
tessl skill review <path>
```

Useful examples:

```bash
tessl skill review ~/.openclaw/skills/meta-cognition
tessl skill review ./skills/work-to-skill
tessl skill review ./skills/some-skill/SKILL.md
```

If the exact CLI surface has drifted, inspect:

```bash
tessl --help
tessl skill --help
```

If Tessl is not installed, either:

1. install it with:

```bash
curl -fsSL https://get.tessl.io | sh
```

2. or use the bundled helper:

```bash
scripts/review.sh <path>
```

The helper script will detect whether Tessl exists, print the install command if missing, and run `tessl skill review <path>` when available.

### 3) Extract a structured scorecard

From Tessl review output, capture at least:
- overall score
- strongest areas
- weakest areas
- trigger/description quality issues
- instruction clarity issues
- missing examples / weak workflow guidance
- context bloat or redundancy risks
- publish-readiness judgment

If Tessl returns category scores, preserve them verbatim where possible.

## Manual fallback rubric

If Tessl cannot be installed or executed, do a manual review using this scoring rubric.

Score each dimension from **1-5**:
- **Trigger clarity**: does the description clearly say what the skill does and when to use it?
- **Workflow executability**: can another agent follow the steps without guessing?
- **Context efficiency**: is the skill lean, or does it waste context?
- **Reusability**: does it avoid hidden tribal knowledge and local-only assumptions?
- **Safety**: does it properly constrain risky or irreversible actions?

Convert to a 100-point score:

```text
Total = (sum of 5 dimension scores / 25) * 100
```

Verdict bands:
- **90-100**: publish-ready
- **75-89**: strong, but improve a few areas
- **60-74**: useful, but needs substantial revision
- **<60**: not ready

Always state clearly whether the score came from **Tessl** or from the **manual fallback rubric**.

## Secondary workflow: scenario-based evaluation

When the user wants deeper validation, go beyond `skill review` and run scenario evals.

Use Tessl scenario tooling when the question is not just “is this well-written?” but “does this skill actually improve agent performance?”

Preferred flow:

```bash
tessl scenario generate <path>
tessl scenario run <path-or-scenario>
```

Use scenario evals for:
- regression checks after editing a skill
- comparing two versions of a skill
- checking whether extra context actually helps
- judging real task success rather than surface quality only

## What to look for in your analysis

After Tessl output, add your own judgment across these dimensions:

### 1. Trigger quality
- Is the frontmatter description specific enough to trigger reliably?
- Does it say both **what the skill does** and **when to use it**?
- Is it too vague, too generic, or too narrow?

### 2. Workflow quality
- Are the steps executable?
- Does the skill guide the agent through decisions, not just dump information?
- Are fragile steps sufficiently constrained?

### 3. Context efficiency
- Is the body concise enough?
- Does it duplicate obvious model knowledge?
- Should detailed material move into references instead of bloating `SKILL.md`?

### 4. Reusability
- Would another agent instance be able to use this without extra tribal knowledge?
- Are assumptions, prerequisites, and inputs explicit?

### 5. Safety and overreach
- Does the skill push the agent toward risky or irreversible actions without proper checks?
- Are approval boundaries and destructive actions handled clearly?

## Output format

Use this output shape unless the user asks for another format:

```markdown
## Tessl Skill Review
- Target:
- Tessl overall score:
- Verdict: ready / close / needs work / not publishable yet

## Strengths
- ...

## Weaknesses
- ...

## High-impact fixes
1. ...
2. ...
3. ...

## Suggested rewrite areas
- frontmatter:
- workflow:
- examples:
- references/scripts:

## Final recommendation
- ...
```

## Publishing / PR use

When reviewing skills for a PR or public registry submission:
- use Tessl score as an input, not the only decision-maker
- call out any mismatch between score and real-world usefulness
- flag private-environment coupling, hardcoded paths, secret handling, or weak public readability

## Anti-patterns

Do not:
- give a score without actually running Tessl when Tessl is available
- confuse “nice writing” with “effective agent behavior”
- accept a high-level skill that has no actionable workflow
- ignore bloated context just because the prose sounds polished
- assume a skill is good only because it is long

## Quick command checklist

```bash
# install if needed
curl -fsSL https://get.tessl.io | sh

# or use the bundled helper
scripts/review.sh <path>

# inspect CLI if unsure
tessl --help
tessl skill --help

# basic quality review
tessl skill review <path>

# deeper evals when needed
tessl scenario generate <path>
tessl scenario run <path-or-scenario>
```

## Trigger phrases

- “测一下这个 skill 的评分”
- “帮我评估这个 SKILL.md”
- “这个 skill 质量怎么样”
- “用 Tessl 跑一下 skill review”
- “compare these two skill versions”
- “is this skill publish-ready?”
- “score this skill”
- “review this skill against best practices”

Related Skills

weekly-review

33
from aAAaqwq/AGI-Super-Team

Weekly project review report

requesting-code-review

33
from aAAaqwq/AGI-Super-Team

Use when completing tasks, implementing major features, or before merging to verify work meets requirements

receiving-code-review

33
from aAAaqwq/AGI-Super-Team

Use when receiving code review feedback, before implementing suggestions, especially if feedback seems unclear or technically questionable - requires technical rigor and verification, not performative agreement or blind implementation

performing-security-code-review

33
from aAAaqwq/AGI-Super-Team

This skill provides automated assistance for security agent tasks Execute this skill enables AI assistant to conduct a security-focused code review using the security-agent plugin. it analyzes code for potential vulnerabilities like sql injection, xss, authentication flaws, and insecure dependencies. AI assistant uses this skill wh... Use when assessing security or running audits. Trigger with phrases like 'security scan', 'audit', or 'vulnerability'.

metrics-review

33
from aAAaqwq/AGI-Super-Team

Review and analyze product metrics with trend analysis and actionable insights. Use when running a weekly, monthly, or quarterly metrics review, investigating a sudden spike or drop, comparing performance against targets, or turning raw numbers into a scorecard with recommended actions.

legal-review

33
from aAAaqwq/AGI-Super-Team

Review legal documents (NDA, contracts, agreements) for sensitive clauses, risks, and red flags

contract-review

33
from aAAaqwq/AGI-Super-Team

Analyze contracts for risks, check completeness, and provide actionable recommendations. Supports employment contracts, NDAs, service agreements, and more.

code-review-quality

33
from aAAaqwq/AGI-Super-Team

Conduct context-driven code reviews focusing on quality, testability, and maintainability. Use when reviewing code, providing feedback, or establishing review practices.

change-review

33
from aAAaqwq/AGI-Super-Team

Validate CRM/PM changes before PR

wemp-operator

33
from aAAaqwq/AGI-Super-Team

> 微信公众号全功能运营——草稿/发布/评论/用户/素材/群发/统计/菜单/二维码 API 封装

Content & Documentation

zsxq-smart-publish

33
from aAAaqwq/AGI-Super-Team

Publish and manage content on 知识星球 (zsxq.com). Supports talk posts, Q&A, long articles, file sharing, digest/bookmark, homework tasks, and tag management. Use when publishing content to 知识星球, creating/editing posts, uploading files/images/audio, managing digests, batch publishing, or formatting content for 知识星球.

zoom-automation

33
from aAAaqwq/AGI-Super-Team

Automate Zoom meeting creation, management, recordings, webinars, and participant tracking via Rube MCP (Composio). Always search tools first for current schemas.