agentic-workflow

Agentic Workflow Pattern

422 stars

Best use case

agentic-workflow is best used when you need a repeatable AI agent workflow instead of a one-off prompt.

Agentic Workflow Pattern

Teams using agentic-workflow should expect a more consistent output, faster repeated execution, less prompt rewriting.

When to use this skill

  • You want a reusable workflow that can be run more than once with consistent structure.

When not to use this skill

  • You only need a quick one-off answer and do not need a reusable workflow.
  • You cannot install or maintain the underlying files, dependencies, or repository context.

Installation

Claude Code / Cursor / Codex

$curl -o ~/.claude/skills/agentic-workflow/SKILL.md --create-dirs "https://raw.githubusercontent.com/vibeeval/vibecosystem/main/skills/agentic-workflow/SKILL.md"

Manual Installation

  1. Download SKILL.md from GitHub
  2. Place it in .claude/skills/agentic-workflow/SKILL.md inside your project
  3. Restart your AI agent — it will auto-discover the skill

How agentic-workflow Compares

Feature / Agentagentic-workflowStandard Approach
Platform SupportNot specifiedLimited / Varies
Context Awareness High Baseline
Installation ComplexityUnknownN/A

Frequently Asked Questions

What does this skill do?

Agentic Workflow Pattern

Where can I find the source code?

You can find the source code on GitHub using the link provided at the top of the page.

Related Guides

SKILL.md Source

# Agentic Workflow Pattern

Standard multi-agent pipeline for implementation tasks.

## Architecture Principles

- Use `run_in_background: true` for all agents to keep main context minimal
- Use `Task` tool (never `TaskOutput`) to avoid receiving full agent transcripts
- Agents write outputs to `.claude/cache/agents/<stage>/` for injection into subsequent agents
- Main conversation is pure orchestration — no heavy lifting, only coordination

## Workflow Stages

### 1. Research Agent
```
Task(subagent_type="oracle", run_in_background=true, prompt="""
Query NIA Oracle (via /nia-docs skill) to verify approach and gather best practices.

Output to: .claude/cache/agents/oracle/<task>-research.md
""")
```
- Enforce NIA as the research layer
- Output: Research findings

### 2. Planning Agent
```
Task(subagent_type="plan-agent", run_in_background=true, prompt="""
Read: .claude/cache/agents/oracle/<task>-research.md
Use RP-CLI to analyze the target codebase section.
Generate implementation plan informed by research.

Output to: .claude/cache/agents/plan-agent/<task>-plan.md
""")
```
- Receives: Research agent output as context
- Output: Implementation plan

### 3. Validation Agent
```
Task(subagent_type="validate-agent", run_in_background=true, prompt="""
Read: .claude/cache/agents/plan-agent/<task>-plan.md
Read: .claude/cache/agents/oracle/<task>-research.md
Review plan against research findings and best practices.

Output to: .claude/cache/agents/validate-agent/<task>-validated.md
""")
```
- Reviews plan against research
- Output: Validated plan with amendments

### 4. Implementation Agent
```
Task(subagent_type="agentica-agent", run_in_background=true, prompt="""
Read: .claude/cache/agents/validate-agent/<task>-validated.md
Read: .claude/cache/agents/oracle/<task>-research.md

TDD approach: Write failing tests FIRST, then implement.
Run tests to verify.

Output summary to: .claude/cache/agents/implement-agent/<task>-implementation.md
""")
```
- Receives: Validated plan + research context
- **TDD**: Failing tests first
- Output: Implementation + tests

### 5. Review Agent
```
Task(subagent_type="review-agent", run_in_background=true, prompt="""
Read: .claude/cache/agents/implement-agent/<task>-implementation.md
Read: .claude/cache/agents/validate-agent/<task>-validated.md
Read: .claude/cache/agents/oracle/<task>-research.md

Cross-reference implementation against plan and research.
Run tests to confirm passing.

Output to: .claude/cache/agents/review-agent/<task>-review.md
""")
```
- Cross-references all artifacts
- Confirms tests pass
- Output: Review summary

## Agent Progress Monitoring

```bash
# Watch for system reminders:
# "Agent a42a16e progress: 6 new tools used, 88914 new tokens"

# Poll for output files:
find .claude/cache/agents -name "*.md" -mmin -5

# Check task file size growth:
wc -c /tmp/claude/.../tasks/<id>.output
```

**Stuck detection:**
1. Progress reminders stop arriving
2. Task output file size stops growing
3. Expected output file not created after reasonable time

## Directory Structure

```
.claude/cache/agents/
├── oracle/
│   └── <task>-research.md
├── plan-agent/
│   └── <task>-plan.md
├── validate-agent/
│   └── <task>-validated.md
├── implement-agent/
│   └── <task>-implementation.md
└── review-agent/
    └── <task>-review.md
```

## Key Rules

1. **Never use TaskOutput** - floods context with 70k+ token transcripts
2. **Always run_in_background=true** - isolates agent context
3. **File-based handoff** - each agent reads previous agent's output file
4. **Poll, don't block** - check file system for outputs, don't wait
5. **TDD in implementation** - failing tests first, then make them pass

## Source
- Session 2026-01-01: SDK Phase 3 implementation using this pattern

Related Skills

workflow-router

422
from vibeeval/vibecosystem

Goal-based workflow orchestration - routes tasks to specialist agents based on user goals

tdd-workflow

422
from vibeeval/vibecosystem

Use this skill when writing new features, fixing bugs, or refactoring code. Enforces test-driven development with 80%+ coverage including unit, integration, and E2E tests.

agentica-spawn

422
from vibeeval/vibecosystem

Spawn Agentica multi-agent patterns

agentica-server

422
from vibeeval/vibecosystem

Agentica server + Claude proxy setup - architecture, startup sequence, debugging

agentica-sdk

422
from vibeeval/vibecosystem

Build Python agents with Agentica SDK - @agentic decorator, spawn(), persistence, MCP integration

agentica-prompts

422
from vibeeval/vibecosystem

Write reliable prompts for Agentica/REPL agents that avoid LLM instruction ambiguity

agentica-infrastructure

422
from vibeeval/vibecosystem

Reference guide for Agentica multi-agent infrastructure APIs

agentica-claude-proxy

422
from vibeeval/vibecosystem

Guide for integrating Agentica SDK with Claude Code CLI proxy

wiring

422
from vibeeval/vibecosystem

Wiring Verification

websocket-patterns

422
from vibeeval/vibecosystem

Connection management, room patterns, reconnection strategies, message buffering, and binary protocol design.

visual-verdict

422
from vibeeval/vibecosystem

Screenshot comparison QA for frontend development. Takes a screenshot of the current implementation, scores it across multiple visual dimensions, and returns a structured PASS/REVISE/FAIL verdict with concrete fixes. Use when implementing UI from a design reference or verifying visual correctness.

verification-loop

422
from vibeeval/vibecosystem

Comprehensive verification system covering build, types, lint, tests, security, and diff review before a PR.