error-debugging-multi-agent-review
Use when working with error debugging multi agent review
About this skill
This AI agent skill, `error-debugging-multi-agent-review`, provides a structured methodology and set of guidelines for agents involved in the orchestration and execution of multi-agent code review tasks, specifically focusing on error debugging. It equips agents with instructions to clarify goals, identify constraints, apply relevant best practices, and validate outcomes when engaging in complex code review workflows involving several AI entities. Part of the "antigravity-awesome-skills" repository, it aims to enhance agents' ability to perform sophisticated development operations by providing a playbook for collaborative error resolution.
Best use case
Guiding AI agents in orchestrating and performing error debugging within collaborative code review environments involving multiple AI agents.
Use when working with error debugging multi agent review
The agent will follow a structured approach to error debugging in multi-agent code reviews, leading to clearer goals, adherence to best practices, actionable steps, and verified results. This should result in more effective and systematic code reviews and error resolution.
Practical example
Example input
Initiate a multi-agent code review for a given feature branch, focusing on identifying and debugging runtime errors in a Python script where another agent has reported an off-by-one error in a loop.
Example output
{
"clarified_goals": "1. Understand the Python script's purpose and the reported off-by-one error. 2. Collaborate with the bug-identifying agent to reproduce the error. 3. Engage a Python best practices agent to propose fixes and review the corrected logic.",
"identified_constraints": "Limited access to the full codebase beyond the provided script. Deadline for review is end of day.",
"best_practices_applied": "1. Step-by-step debugging. 2. Unit test creation for the identified bug. 3. Peer review (agent-to-agent) of proposed changes. 4. Code quality checks.",
"actionable_steps": [
"Agent 1 (current): Request the full context of the reported off-by-one error from the identifying agent.",
"Agent 1 (current): Create a minimal reproducible example (MRE) if not already provided.",
"Agent 2 (Python expert): Analyze the MRE and suggest potential fixes for the loop logic.",
"Agent 1 (current): Apply the suggested fix in a test environment.",
"Agent 1 (current): Verify the fix using the MRE and any new unit tests.",
"Agent 3 (Reviewer): Review the proposed fix and its verification for correctness and adherence to Python best practices."
],
"verification_outcome": "The off-by-one error has been reproduced, a fix has been proposed and implemented, and verification tests confirm the bug is resolved without introducing new issues. The code adheres to Python best practices."
}When to use this skill
- Working on tasks related to orchestrating multi-agent code reviews.
- Needing best practices, guidance, or checklists for multi-agent code review processes.
- When the agent needs to debug errors collaboratively in a code review context.
When not to use this skill
- The task is unrelated to code review or multi-agent orchestration.
- A different domain or tool is required outside the scope of multi-agent code review.
- The task involves a single-agent code review without the need for debugging orchestration.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/error-debugging-multi-agent-review/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How error-debugging-multi-agent-review Compares
| Feature / Agent | error-debugging-multi-agent-review | Standard Approach |
|---|---|---|
| Platform Support | Claude | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | medium | N/A |
Frequently Asked Questions
What does this skill do?
Use when working with error debugging multi agent review
Which AI agents support this skill?
This skill is designed for Claude.
How difficult is it to install?
The installation complexity is rated as medium. You can find the installation instructions above.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
Related Guides
AI Agents for Coding
Browse AI agent skills for coding, debugging, testing, refactoring, code review, and developer workflows across Claude, Cursor, and Codex.
Best AI Skills for Claude
Explore the best AI skills for Claude and Claude Code across coding, research, workflow automation, documentation, and agent operations.
Cursor vs Codex for AI Workflows
Compare Cursor and Codex for AI coding workflows, repository assistance, debugging, refactoring, and reusable developer skills.
SKILL.md Source
# Multi-Agent Code Review Orchestration Tool
## Use this skill when
- Working on multi-agent code review orchestration tool tasks or workflows
- Needing guidance, best practices, or checklists for multi-agent code review orchestration tool
## Do not use this skill when
- The task is unrelated to multi-agent code review orchestration tool
- You need a different domain or tool outside this scope
## Instructions
- Clarify goals, constraints, and required inputs.
- Apply relevant best practices and validate outcomes.
- Provide actionable steps and verification.
- If detailed examples are required, open `resources/implementation-playbook.md`.
## Role: Expert Multi-Agent Review Orchestration Specialist
A sophisticated AI-powered code review system designed to provide comprehensive, multi-perspective analysis of software artifacts through intelligent agent coordination and specialized domain expertise.
## Context and Purpose
The Multi-Agent Review Tool leverages a distributed, specialized agent network to perform holistic code assessments that transcend traditional single-perspective review approaches. By coordinating agents with distinct expertise, we generate a comprehensive evaluation that captures nuanced insights across multiple critical dimensions:
- **Depth**: Specialized agents dive deep into specific domains
- **Breadth**: Parallel processing enables comprehensive coverage
- **Intelligence**: Context-aware routing and intelligent synthesis
- **Adaptability**: Dynamic agent selection based on code characteristics
## Tool Arguments and Configuration
### Input Parameters
- `$ARGUMENTS`: Target code/project for review
- Supports: File paths, Git repositories, code snippets
- Handles multiple input formats
- Enables context extraction and agent routing
### Agent Types
1. Code Quality Reviewers
2. Security Auditors
3. Architecture Specialists
4. Performance Analysts
5. Compliance Validators
6. Best Practices Experts
## Multi-Agent Coordination Strategy
### 1. Agent Selection and Routing Logic
- **Dynamic Agent Matching**:
- Analyze input characteristics
- Select most appropriate agent types
- Configure specialized sub-agents dynamically
- **Expertise Routing**:
```python
def route_agents(code_context):
agents = []
if is_web_application(code_context):
agents.extend([
"security-auditor",
"web-architecture-reviewer"
])
if is_performance_critical(code_context):
agents.append("performance-analyst")
return agents
```
### 2. Context Management and State Passing
- **Contextual Intelligence**:
- Maintain shared context across agent interactions
- Pass refined insights between agents
- Support incremental review refinement
- **Context Propagation Model**:
```python
class ReviewContext:
def __init__(self, target, metadata):
self.target = target
self.metadata = metadata
self.agent_insights = {}
def update_insights(self, agent_type, insights):
self.agent_insights[agent_type] = insights
```
### 3. Parallel vs Sequential Execution
- **Hybrid Execution Strategy**:
- Parallel execution for independent reviews
- Sequential processing for dependent insights
- Intelligent timeout and fallback mechanisms
- **Execution Flow**:
```python
def execute_review(review_context):
# Parallel independent agents
parallel_agents = [
"code-quality-reviewer",
"security-auditor"
]
# Sequential dependent agents
sequential_agents = [
"architecture-reviewer",
"performance-optimizer"
]
```
### 4. Result Aggregation and Synthesis
- **Intelligent Consolidation**:
- Merge insights from multiple agents
- Resolve conflicting recommendations
- Generate unified, prioritized report
- **Synthesis Algorithm**:
```python
def synthesize_review_insights(agent_results):
consolidated_report = {
"critical_issues": [],
"important_issues": [],
"improvement_suggestions": []
}
# Intelligent merging logic
return consolidated_report
```
### 5. Conflict Resolution Mechanism
- **Smart Conflict Handling**:
- Detect contradictory agent recommendations
- Apply weighted scoring
- Escalate complex conflicts
- **Resolution Strategy**:
```python
def resolve_conflicts(agent_insights):
conflict_resolver = ConflictResolutionEngine()
return conflict_resolver.process(agent_insights)
```
### 6. Performance Optimization
- **Efficiency Techniques**:
- Minimal redundant processing
- Cached intermediate results
- Adaptive agent resource allocation
- **Optimization Approach**:
```python
def optimize_review_process(review_context):
return ReviewOptimizer.allocate_resources(review_context)
```
### 7. Quality Validation Framework
- **Comprehensive Validation**:
- Cross-agent result verification
- Statistical confidence scoring
- Continuous learning and improvement
- **Validation Process**:
```python
def validate_review_quality(review_results):
quality_score = QualityScoreCalculator.compute(review_results)
return quality_score > QUALITY_THRESHOLD
```
## Example Implementations
### 1. Parallel Code Review Scenario
```python
multi_agent_review(
target="/path/to/project",
agents=[
{"type": "security-auditor", "weight": 0.3},
{"type": "architecture-reviewer", "weight": 0.3},
{"type": "performance-analyst", "weight": 0.2}
]
)
```
### 2. Sequential Workflow
```python
sequential_review_workflow = [
{"phase": "design-review", "agent": "architect-reviewer"},
{"phase": "implementation-review", "agent": "code-quality-reviewer"},
{"phase": "testing-review", "agent": "test-coverage-analyst"},
{"phase": "deployment-readiness", "agent": "devops-validator"}
]
```
### 3. Hybrid Orchestration
```python
hybrid_review_strategy = {
"parallel_agents": ["security", "performance"],
"sequential_agents": ["architecture", "compliance"]
}
```
## Reference Implementations
1. **Web Application Security Review**
2. **Microservices Architecture Validation**
## Best Practices and Considerations
- Maintain agent independence
- Implement robust error handling
- Use probabilistic routing
- Support incremental reviews
- Ensure privacy and security
## Extensibility
The tool is designed with a plugin-based architecture, allowing easy addition of new agent types and review strategies.
## Invocation
Target for review: $ARGUMENTSRelated Skills
code-review-excellence
Transform code reviews from gatekeeping to knowledge sharing through constructive feedback, systematic analysis, and collaborative improvement.
find-bugs
Find bugs, security vulnerabilities, and code quality issues in local branch changes. Use when asked to review changes, find bugs, security review, or audit code on the current branch.
multi-platform-apps-multi-platform
Build and deploy the same feature consistently across web, mobile, and desktop platforms using API-first architecture and parallel implementation strategies.
multi-cloud-architecture
Decision framework and patterns for architecting applications across AWS, Azure, and GCP.
multi-agent-brainstorming
Simulate a structured peer-review process using multiple specialized agents to validate designs, surface hidden assumptions, and identify failure modes before implementation.
multi-advisor
Conselho de especialistas — consulta multiplos agentes do ecossistema em paralelo para analise multi-perspectiva de qualquer topico. Ativa personas, especialistas e agentes tecnicos simultaneamente, cada um pela sua otica unica, e consolida em sintese decisoria final.
lightning-architecture-review
Review Bitcoin Lightning Network protocol designs, compare channel factory approaches, and analyze Layer 2 scaling tradeoffs. Covers trust models, on-chain footprint, consensus requirements, HTLC/PTLC compatibility, liveness, and watchtower support.
gha-security-review
Find exploitable vulnerabilities in GitHub Actions workflows. Every finding MUST include a concrete exploitation scenario — if you can't build the attack, don't report it.
gh-review-requests
Fetch unread GitHub notifications for open PRs where review is requested from a specified team or opened by a team member. Use when asked to "find PRs I need to review", "show my review requests", "what needs my review", "fetch GitHub review requests", or "check team review queue".
fix-review
Verify fix commits address audit findings without new bugs
error-diagnostics-smart-debug
Use when working with error diagnostics smart debug
error-diagnostics-error-trace
You are an error tracking and observability expert specializing in implementing comprehensive error monitoring solutions. Set up error tracking systems, configure alerts, implement structured logging,