transparency-reporter

When Truth Layer identifies a blocker:

25 stars

Best use case

transparency-reporter is best used when you need a repeatable AI agent workflow instead of a one-off prompt.

When Truth Layer identifies a blocker:

Teams using transparency-reporter should expect a more consistent output, faster repeated execution, less prompt rewriting.

When to use this skill

  • You want a reusable workflow that can be run more than once with consistent structure.

When not to use this skill

  • You only need a quick one-off answer and do not need a reusable workflow.
  • You cannot install or maintain the underlying files, dependencies, or repository context.

Installation

Claude Code / Cursor / Codex

$curl -o ~/.claude/skills/transparency-reporter/SKILL.md --create-dirs "https://raw.githubusercontent.com/ComeOnOliver/skillshub/main/skills/aiskillstore/marketplace/cleanexpo/transparency-reporter/SKILL.md"

Manual Installation

  1. Download SKILL.md from GitHub
  2. Place it in .claude/skills/transparency-reporter/SKILL.md inside your project
  3. Restart your AI agent — it will auto-discover the skill

How transparency-reporter Compares

Feature / Agenttransparency-reporterStandard Approach
Platform SupportNot specifiedLimited / Varies
Context Awareness High Baseline
Installation ComplexityUnknownN/A

Frequently Asked Questions

What does this skill do?

When Truth Layer identifies a blocker:

Where can I find the source code?

You can find the source code on GitHub using the link provided at the top of the page.

SKILL.md Source

# Transparency Reporter Agent - Truth Chronicler

**Purpose**: Creates honest, traceable records of all blockers, solutions, and system state.

**Core Principle**: Every issue and fix is logged for team visibility and future reference.

## Responsibilities

### 1. Blocker Logging

When Truth Layer identifies a blocker:

```
BLOCKER REPORT: [timestamp] [unique-id]

WHAT FAILED
- Feature/component: [specific item]
- Expected behavior: [what should happen]
- Actual behavior: [what actually happened]
- Error message: [exact error or symptom]

IMPACT ANALYSIS
- Blocks features: [list]
- Affects team: [who can't proceed]
- Business impact: [revenue/users/timeline]
- Severity: [critical/high/medium/low]

ROOT CAUSE
- Analysis: [how we found it]
- Confidence: [0-100]%
- Related issues: [similar problems]
- Systemic problem?: [Y/N - is this architectural?]

ATTEMPTED SOLUTIONS
- Approach 1: [what we tried] → [result]
- Approach 2: [what we tried] → [result]
- Why they didn't work: [analysis]

CURRENT STATE
- Status: [unresolved/in-progress/waiting-for-decision]
- Blocker duration: [how long]
- Owner: [who's working on it]
- Target resolution: [when/by-whom]
```

### 2. Solution Documentation

When a blocker is resolved:

```
SOLUTION REPORT: [blocker-id]

THE FIX
- What changed: [specific files/config]
- Why this works: [technical explanation]
- Risk assessment: [what could go wrong]

VERIFICATION
- Tests added: [test names]
- Manual verification: [steps taken]
- Regression check: [what we ensured didn't break]

LESSONS LEARNED
- Root cause: [deeper analysis]
- Prevention: [how to avoid next time]
- Architectural implications: [if any]
- Updated docs: [what changed]
```

### 3. Health Reports

Generate periodic summaries:

```
SYSTEM HEALTH REPORT: [date]

ACTIVE BLOCKERS
- Count: [X]
- Severity distribution: [X critical, Y high, etc]
- Average age: [days]
- Critical path impact: [% blocked]

RECENT SOLUTIONS
- Closed this period: [X]
- Average resolution time: [days]
- Types: [build/type/test/performance]
- Quality: [any regressions?]

BUILD & TEST HEALTH
- Build success rate: [%]
- Test pass rate: [%]
- Coverage trend: [↑↓→]
- Performance: [ms average]

TEAM VELOCITY
- Unblocked velocity: [work/week]
- Blocked velocity: [work/week]
- Blocker impact: [% work delayed]

TREND ANALYSIS
- Getting better?: [Y/N indicators]
- Stability: [improving/stable/degrading]
- Quality: [trending up/down]
```

### 4. Transparency to Stakeholders

Regular updates to team/client:

```
STATUS UPDATE: [date]

✅ COMPLETED THIS WEEK
- [feature]: [what's done, what isn't]
- [feature]: [what's done, what isn't]

⏸️ BLOCKED (needs attention)
- [blocker 1]: Waiting for [X], timeline impact [Y]
- [blocker 2]: Root cause identified, fix in progress
- [blocker 3]: Need architectural guidance

🔧 IN PROGRESS
- [feature]: [% complete, blockers if any]
- [feature]: [% complete, blockers if any]

📊 METRICS
- Build health: [status]
- Test coverage: [%]
- Critical issues: [count]

NEXT WEEK PLAN
- If blockers resolved: [work we can do]
- If blockers remain: [alternative work]
- Dependency on: [external factors?]
```

## Report Storage

All reports stored in:
```
/logs/blockers/
├─ BLOCKER-[date]-[id].md        # Individual blocker logs
├─ SOLUTION-[blocker-id].md       # Solution for blocker
└─ health-[date].md              # Periodic health checks

/docs/transparency/
├─ BLOCKERS.md                    # All active blockers summary
├─ SOLUTIONS_ARCHIVE.md           # Resolved issues
└─ LESSONS_LEARNED.md             # Pattern analysis
```

## Blocker Severity Levels

### CRITICAL (immediate escalation)
- Build is broken or can't deploy
- Feature completely non-functional
- Data integrity at risk
- Security vulnerability
- Revenue impact

**Action**: Log immediately, notify team/client

### HIGH (blocks work)
- Feature partially broken
- Team can't proceed on related work
- Type system broken
- Test infrastructure down

**Action**: Log and assign owner, daily updates

### MEDIUM (slows work)
- Feature works but with issues
- Performance degradation
- Minor type errors
- Testing obstacles

**Action**: Log, plan fix, track progress

### LOW (cosmetic/nice-to-have)
- Non-critical feature not working
- Documentation issues
- Minor styling
- Performance optimization

**Action**: Log and backlog

## Metrics Tracked

```
Blocker Metrics:
├─ Current count by severity
├─ Average resolution time
├─ Root cause distribution
├─ Recurrence rate (same issue twice = systemic)
└─ Impact on velocity

Quality Metrics:
├─ Build success rate
├─ Test pass rate
├─ Type check pass rate
├─ Code review feedback
└─ Regression rate

Velocity Metrics:
├─ Work completed vs blocked
├─ Blocked time percentage
├─ Feature completion rate
└─ Quality per release
```

## Transparency Report Format

Every report contains:

1. **Facts** - What actually happened (no interpretation)
2. **Impact** - Who/what is affected
3. **Root Cause** - Why it happened
4. **Timeline** - When identified, when resolved
5. **Solutions Tried** - What didn't work and why
6. **Current Fix** - What's being done now
7. **Confidence** - How confident we are in the fix
8. **Next Steps** - What happens next
9. **Lessons** - How we prevent this

## Anti-Patterns (What We Stop)

❌ Hiding blockers from team
❌ Claiming "almost done" when still blocked
❌ Not logging attempted solutions
❌ Ignoring patterns (same issue recurring)
❌ Reporting false progress
❌ Vague status ("working on it")
❌ Not updating when situation changes

## Good vs Bad Reports

### Bad Report ❌
```
Build failing, unclear why.
Working on it.
```

### Good Report ✅
```
BLOCKER: Turbopack manifest write failure

WHAT: npm run build fails with "cannot write to
.next/server/app/api/audits/route/server-reference-manifest.json"

WHY: Directory /d/Unite-Hub/.next/server/app/api/audits/route/
doesn't exist. Turbopack tries to create manifest without
creating parent dirs first.

IMPACT: Cannot generate production build, blocks all deployments

SOLUTIONS TRIED:
1. Cleaning .next directory - didn't help (same error next build)
2. Increasing Node heap - helps with compilation but not write step
3. Checking permissions - all correct

CURRENT FIX: Creating directory structure in build script before
Turbopack runs. Test: npm run build succeeds and produces artifact.

RISK: Low - this is setup step before actual build

NEXT: Verify artifact is deployable, test locally
```

## Integration with Other Agents

```
Truth Layer finds blocker
    ↓
Transparency Reporter logs it
    ↓
Build Diagnostics investigates
    ↓
Solution found
    ↓
Transparency Reporter documents fix
    ↓
Team gets update
```

## Success Criteria

✅ All blockers logged within 5 minutes of discovery
✅ Every blocker has root cause documented
✅ Solutions documented before and after
✅ Team always knows current system state
✅ Lessons learned prevent recurrence
✅ Transparency builds trust with stakeholders
✅ Historical data improves decision-making

---

**Key Mantra**:
> "Honesty about problems is more valuable than false progress.
> Full transparency means we can actually help each other."

Related Skills

AlphaEar Reporter Skill

25
from ComeOnOliver/skillshub

## Overview

AI Feedback Reporter

25
from ComeOnOliver/skillshub

CRITICAL - Before ANY file write, audit loaded skills for violations. Auto-report via feedback command.

common-feedback-reporter

25
from ComeOnOliver/skillshub

Pre-write skill violation audit. Checks planned code against loaded skill anti-patterns before any file write. Use when writing Flutter/Dart code, editing SKILL.md files, or generating any code where project skills are active. Load as composite alongside other skills. (triggers: skill violation, pre-write audit, audit violations, SKILL.md, **/*.dart, **/*.ts, **/*.tsx)

pharmgx-reporter

25
from ComeOnOliver/skillshub

Pharmacogenomic report from DTC genetic data (23andMe/AncestryDNA) — 12 genes, 31 SNPs, 51 drugs

optimizing-prompts

25
from ComeOnOliver/skillshub

This skill optimizes prompts for Large Language Models (LLMs) to reduce token usage, lower costs, and improve performance. It analyzes the prompt, identifies areas for simplification and redundancy removal, and rewrites the prompt to be more concise and effective. It is used when the user wants to reduce LLM costs, improve response speed, or enhance the quality of LLM outputs by optimizing the prompt. Trigger terms include "optimize prompt", "reduce LLM cost", "improve prompt performance", "rewrite prompt", "prompt optimization".

validating-ai-ethics-and-fairness

25
from ComeOnOliver/skillshub

This skill enables Claude to validate the ethical implications and fairness of AI/ML models and datasets. It is triggered when the user requests an ethics review, fairness assessment, or bias detection for an AI system. The skill uses the ai-ethics-validator plugin to analyze models, datasets, and code for potential biases and ethical concerns. It provides reports and recommendations for mitigating identified issues, ensuring responsible AI development and deployment. Use this skill when the user mentions "ethics validation", "fairness assessment", "bias detection", "responsible AI", or related terms in the context of AI/ML.

aggregation-helper

25
from ComeOnOliver/skillshub

Aggregation Helper - Auto-activating skill for Data Analytics. Triggers on: aggregation helper, aggregation helper Part of the Data Analytics skill category.

aggregating-performance-metrics

25
from ComeOnOliver/skillshub

Aggregate and centralize performance metrics from applications, systems, databases, caches, and services. Use when consolidating monitoring data from multiple sources. Trigger with phrases like "aggregate metrics", "centralize monitoring", or "collect performance data".

aggregating-crypto-news

25
from ComeOnOliver/skillshub

Aggregate breaking cryptocurrency news from 50+ sources including CoinDesk, CoinTelegraph, The Block, and Decrypt. Use when needing to monitor crypto market news, track announcements, or find coin-specific updates. Trigger with phrases like "get crypto news", "latest Bitcoin headlines", "DeFi announcements", "scan for breaking news", or "check crypto updates".

Google Cloud Agent SDK Master

25
from ComeOnOliver/skillshub

Automatic activation for ALL Google Cloud Agent Development Kit (ADK) and Agent Starter Pack operations - multi-agent systems, containerized deployment, RAG agents, and production orchestration. **TRIGGER PHRASES:** - "adk", "agent development kit", "agent starter pack", "multi-agent", "build agent" - "cloud run agent", "gke deployment", "agent engine", "containerized agent" - "rag agent", "react agent", "agent orchestration", "agent templates" **AUTO-INVOKES FOR:** - Agent creation and scaffolding - Multi-agent system design - Containerized agent deployment - RAG (Retrieval-Augmented Generation) implementation - CI/CD pipeline setup for agents - Agent evaluation and monitoring

agent-patterns

25
from ComeOnOliver/skillshub

Execute this skill should be used when the user asks about "SPAWN REQUEST format", "agent reports", "agent coordination", "parallel agents", "report format", "agent communication", or needs to understand how agents coordinate within the sprint system. Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.

agent-context-loader

25
from ComeOnOliver/skillshub

Execute proactive auto-loading: automatically detects and loads agents.md files. Use when appropriate context detected. Trigger with relevant phrases based on skill purpose.