code-review-digest-writer
Generates weekly code-review digest docs from PR review comments for any GitHub repository. If present, follows project-specific docs/review-digests/AGENTS.md guidelines. Use this to turn a date-bounded set of PR reviews into a structured markdown “newsletter” that captures themes, repeated issues, and concrete takeaways.
Best use case
code-review-digest-writer is best used when you need a repeatable AI agent workflow instead of a one-off prompt. It is especially useful for teams working in multi. Generates weekly code-review digest docs from PR review comments for any GitHub repository. If present, follows project-specific docs/review-digests/AGENTS.md guidelines. Use this to turn a date-bounded set of PR reviews into a structured markdown “newsletter” that captures themes, repeated issues, and concrete takeaways.
Generates weekly code-review digest docs from PR review comments for any GitHub repository. If present, follows project-specific docs/review-digests/AGENTS.md guidelines. Use this to turn a date-bounded set of PR reviews into a structured markdown “newsletter” that captures themes, repeated issues, and concrete takeaways.
Users should expect a more consistent workflow output, faster repeated execution, and less time spent rewriting prompts from scratch.
Practical example
Example input
Use the "code-review-digest-writer" skill to help with this workflow task. Context: Generates weekly code-review digest docs from PR review comments for any GitHub repository. If present, follows project-specific docs/review-digests/AGENTS.md guidelines. Use this to turn a date-bounded set of PR reviews into a structured markdown “newsletter” that captures themes, repeated issues, and concrete takeaways.
Example output
A structured workflow result with clearer steps, more consistent formatting, and an output that is easier to reuse in the next run.
When to use this skill
- Use this skill when you want a reusable workflow rather than writing the same prompt again and again.
When not to use this skill
- Do not use this when you only need a one-off answer and do not need a reusable workflow.
- Do not use it if you cannot install or maintain the related files, repository context, or supporting tools.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/code-review-digest-writer/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How code-review-digest-writer Compares
| Feature / Agent | code-review-digest-writer | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Generates weekly code-review digest docs from PR review comments for any GitHub repository. If present, follows project-specific docs/review-digests/AGENTS.md guidelines. Use this to turn a date-bounded set of PR reviews into a structured markdown “newsletter” that captures themes, repeated issues, and concrete takeaways.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
Related Guides
AI Agents for Coding
Browse AI agent skills for coding, debugging, testing, refactoring, code review, and developer workflows across Claude, Cursor, and Codex.
Best AI Skills for Claude
Explore the best AI skills for Claude and Claude Code across coding, research, workflow automation, documentation, and agent operations.
Cursor vs Codex for AI Workflows
Compare Cursor and Codex for AI coding workflows, repository assistance, debugging, refactoring, and reusable developer skills.
SKILL.md Source
# Code Review Digest Writer Skill
## When to Use This Skill
- You want a **weekly (or custom window) code-review digest** for the current
repository or project, based on PR review comments.
- You have a **start date** and **end date** and want a markdown newsletter
summarizing what was taught in PR review feedback in that period.
- You want to highlight **themes, repeated issues, and concrete best practices**
rather than just listing PRs.
If the user does not provide both a start and end date, ask them to specify:
`YYYY-MM-DD` → `YYYY-MM-DD` before proceeding.
## Example Prompts
- “Generate a code review digest for this repo for 2025-02-01 → 2025-02-14.”
- “Create a weekly PR review digest for 2025-03-10 → 2025-03-17, using any existing digests as historical context.”
- “From 2025-04-01 to 2025-04-30, summarize what reviewers focused on in `owner/service-repo` and highlight repeated issues.”
- “Write a newsletter-style code review digest for this project for 2025-05-15 → 2025-05-29, tagging themes as [NEW] or [REPEAT].”
## Scope & Repositories
- Target repo: whichever project you want a digest for (typically the repo you
currently have open in Claude Code).
- Ideally the repo contains `docs/review-digests/AGENTS.md` with
project-specific digest guidelines. If not, fall back to the generic structure
described in this Skill and any existing digest files.
- This Skill **must only modify docs** under:
- `docs/review-digests/YYYY-MM-DD.md`
where the date is the **end date** of the reporting window.
- Do **not** modify application code, tests, or configs in the target repo while
this Skill is active. All changes should be within `docs/review-digests/`.
- If `docs/review-digests/` does not exist yet, create the directory before
writing the digest file, so future digests can be added and past ones read.
## Required Local Tools & Assumptions
When using `Bash`, assume:
- `gh` (GitHub CLI) is installed and authenticated for the target repository
(or for a default GitHub identity that can see it).
- Current working directory is the target repo root (or pass `--repo owner/name`
explicitly to `gh` commands, if needed).
If `gh` is not available, gracefully fall back to:
- Summarizing based on locally available PR notes or docs, and
- Clearly stating in the digest that it was generated with partial data.
## High-Level Workflow
When asked to generate or update a digest, follow this workflow:
1. **Confirm time window**
- Ensure you have `start_date` and `end_date` (inclusive).
- Confirm with the user if there is any ambiguity.
2. **Load local digest guidelines**
- If it exists, open `docs/review-digests/AGENTS.md` and read it carefully.
- When present, treat that file as the **source of truth** for:
- What the digest is.
- Where it should be written.
- Required structure and link style.
- How to detect and label repeated issues.
- If it does not exist, follow the layout described later in this Skill and
use any existing digest files in `docs/review-digests/` as a reference.
3. **Inspect existing digests**
- Ensure the `docs/review-digests/` directory exists:
- If it does not, create it; in that case there will be no past digests yet.
- Use `Glob` or `Bash` to list `docs/review-digests/*.md`.
- Load at least the **last 3–4 digests** (if present).
- Extract their **themes and repeated issues** (e.g., fixture reuse,
blind-index invariants, Query() regression patterns, etc.).
- You will use these to detect when issues are recurring.
4. **Fetch PR and review data for the window**
- Use `Bash` with `gh` to query PRs whose:
- `createdAt` is between `[start_date, end_date]`, OR
- `closedAt` is between `[start_date, end_date]`.
- Deduplicate PR numbers.
- For each selected PR:
- Fetch **top-level comments** (PR discussion).
- Fetch **review-thread comments with code context** (via GraphQL).
- Prefer comments from:
- Human reviewers (`__typename == "User"`).
- AI reviewers that contain substantial review content
(e.g., Claude, Copilot PR reviewer).
- Exclude noisy infrastructure/bot comments with no review content
(e.g., `github-actions`, log-only bots, CI status updates).
5. **Cluster feedback into themes**
- Read comments and diff context enough to understand:
- What behavior or pattern was being discussed.
- What best practice or correction was suggested.
- Group comments across PRs into **themes**, such as:
- Logging, Sentry, and performance instrumentation.
- Tests, fixtures, and code structure.
- Security, access control, and PII handling.
- Domain-specific design and invariants for this repository.
- Migrations & tooling.
- Process and meta-patterns in reviews.
6. **Detect repeated issues**
- For each current theme, compare it conceptually to themes you extracted
from previous digests (step 3).
- If the same pattern appears again (e.g., “use TypedDict instead of
`dict[str, Any]` in payloads”, “avoid Django Ninja `Query()` constants”,
“reuse shared fixtures instead of copy-paste”), treat it as a **repeated issue**.
- Use labels:
- `[NEW]` for themes that appear for the first time.
- `[REPEAT]` for themes that have appeared in previous digests.
7. **Draft the digest file**
- Target path: `docs/review-digests/END_DATE.md`
- Example: period `2025-11-13` → `2025-11-27` → `docs/review-digests/2025-11-27.md`.
- Follow the **layout described in `docs/review-digests/AGENTS.md`** and the
most recent digest, including:
- Title with repo and period.
- **Overview** section with 3–6 bullets summarizing main themes.
- Thematic sections (numbered) that group related feedback.
- A closing section (e.g., “How to Use This Digest”).
- Within each section:
- Explain the practice in **plain language**.
- Include 1–3 concrete, generalized examples.
- Call out whether this is `[NEW]` or `[REPEAT]`.
- Emphasize the “why” (business impact, correctness, safety, DX).
8. **Linking to PRs and comments**
- In the body of the digest, use **reference-style links** only:
- `[#2519 – Fix Teams Start Survey race condition][pr-2519]`
- `Key comment: [Fixture reuse recommendation][c-2519-3]`
- At the **bottom of the file**, define every link once:
- `pr-<number>` for PRs.
- `c-<number>-<n>` for specific review comments.
- `ic-<number>-<n>` for issue comments, if needed.
- Reuse identifiers consistently when the same comment is referenced in
multiple sections.
9. **Respect tone and intent**
- The digest is a **newsletter**, not a blame report.
- Highlight:
- What the team is learning.
- Where we’re improving.
- Where patterns are still repeating and need attention.
- Make guidance actionable (e.g., “When adding a new CSV mapping endpoint,
always run through the project’s PII and security checklist docs.”).
10. **Save and review**
- Use `Edit` to create or update the digest file for `END_DATE`.
- Re-open the file after writing to sanity-check:
- Structure matches the prior digests.
- Links resolve correctly and have definitions at the bottom.
- `[NEW]` / `[REPEAT]` tags are applied consistently.
- No accidental code changes occurred in the repo.
## Output Expectations
When this Skill is active and asked to generate a digest, your final output
should be:
- A **single markdown file** under `docs/review-digests/YYYY-MM-DD.md`.
- A short natural-language summary back to the user describing:
- The period covered.
- The main themes identified.
- How many themes were `[REPEAT]` vs `[NEW]`.
If you were unable to access GitHub or some PRs, clearly note in the digest and
in your summary which data sources were missing and how that might limit the
digest.
## Severity / Emphasis Tags
Instead of issue severities, this Skill uses **learning/emphasis tags**:
- `[NEW]` – First time this theme appears in digests.
- `[REPEAT]` – Theme appeared in at least one prior digest.
- `[HIGH-IMPACT]` – Optional extra tag for themes with clear business impact
(e.g., security invariants, multi-tenant correctness, high-risk migrations).
Use these tags sparingly and consistently; they should help readers prioritize
which lessons to internalize first.
## Compatibility Notes
This skill is designed to work with both **Claude Code** and **OpenAI Codex**.
For Codex users:
- Install via skill-installer with `--repo DiversioTeam/agent-skills-marketplace
--path plugins/code-review-digest-writer/skills/code-review-digest-writer`.
- Use `$skill code-review-digest-writer` to invoke.
For Claude Code users:
- Install via `/plugin install code-review-digest-writer@diversiotech`.
- Use `/code-review-digest-writer:review-digest` to invoke.Related Skills
prd-doc-writer
Write and iteratively refine PRD/需求文档 with a story-driven structure and strict staged confirmations (journey map alignment, per-story single-point confirmation, final generation gate). Use when the user asks to 梳理/撰写/完善 PRD、需求文档、用户故事、验收标准,并希望用 ASCII 线框图与 Mermaid(流程图/状态图/时序图)来减少歧义、共同完成文档。
woocommerce-code-review
Review WooCommerce code changes for coding standards compliance. Use when reviewing code locally, performing automated PR reviews, or checking code quality.
svelte-code-writer
CLI tools for Svelte 5 documentation lookup and code analysis. MUST be used whenever creating, editing or analyzing any Svelte component (.svelte) or Svelte module (.svelte.ts/.svelte.js). If possible, this skill should be executed within the svelte-file-editor agent for optimal results.
writer
Document creation, format conversion (ODT/DOCX/PDF), mail merge, and automation with LibreOffice Writer.
wiki-page-writer
Generates rich technical documentation pages with dark-mode Mermaid diagrams, source code citations, and first-principles depth. Use when writing documentation, generating wiki pages, creating technical deep-dives, or documenting specific components or systems.
seo-content-writer
Writes SEO-optimized content based on provided keywords and topic briefs. Creates engaging, comprehensive content following best practices. Use PROACTIVELY for content creation tasks.
security-review
Use this skill when adding authentication, handling user input, working with secrets, creating API endpoints, or implementing payment/sensitive features. Provides comprehensive security checklist and patterns.
performance-testing-review-multi-agent-review
Use when working with performance testing review multi agent review
performance-testing-review-ai-review
You are an expert AI-powered code review specialist combining automated static analysis, intelligent pattern recognition, and modern DevOps practices. Leverage AI tools (GitHub Copilot, Qodo, GPT-5, C
fix-review
Verify fix commits address audit findings without new bugs
error-debugging-multi-agent-review
Use when working with error debugging multi agent review
comprehensive-review-pr-enhance
You are a PR optimization expert specializing in creating high-quality pull requests that facilitate efficient code reviews. Generate comprehensive PR descriptions, automate review processes, and ensure PRs follow best practices for clarity, size, and reviewability.