qa-expert
This skill should be used when establishing comprehensive QA testing processes for any software project. Use when creating test strategies, writing test cases following Google Testing Standards, executing test plans, tracking bugs with P0-P4 classification, calculating quality metrics, or generating progress reports. Includes autonomous execution capability via master prompts and complete documentation templates for third-party QA team handoffs. Implements OWASP security testing and achieves 90% coverage targets.
Best use case
qa-expert is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
This skill should be used when establishing comprehensive QA testing processes for any software project. Use when creating test strategies, writing test cases following Google Testing Standards, executing test plans, tracking bugs with P0-P4 classification, calculating quality metrics, or generating progress reports. Includes autonomous execution capability via master prompts and complete documentation templates for third-party QA team handoffs. Implements OWASP security testing and achieves 90% coverage targets.
Teams using qa-expert should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/qa-expert/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How qa-expert Compares
| Feature / Agent | qa-expert | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
This skill should be used when establishing comprehensive QA testing processes for any software project. Use when creating test strategies, writing test cases following Google Testing Standards, executing test plans, tracking bugs with P0-P4 classification, calculating quality metrics, or generating progress reports. Includes autonomous execution capability via master prompts and complete documentation templates for third-party QA team handoffs. Implements OWASP security testing and achieves 90% coverage targets.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# QA Expert Establish world-class QA testing processes for any software project using proven methodologies from Google Testing Standards and OWASP security best practices. ## When to Use This Skill Trigger this skill when: - Setting up QA infrastructure for a new or existing project - Writing standardized test cases (AAA pattern compliance) - Executing comprehensive test plans with progress tracking - Implementing security testing (OWASP Top 10) - Filing bugs with proper severity classification (P0-P4) - Generating QA reports (daily summaries, weekly progress) - Calculating quality metrics (pass rate, coverage, gates) - Preparing QA documentation for third-party team handoffs - Enabling autonomous LLM-driven test execution ## Quick Start **One-command initialization**: ```bash python scripts/init_qa_project.py <project-name> [output-directory] ``` **What gets created**: - Directory structure (`tests/docs/`, `tests/e2e/`, `tests/fixtures/`) - Tracking CSVs (`TEST-EXECUTION-TRACKING.csv`, `BUG-TRACKING-TEMPLATE.csv`) - Documentation templates (`BASELINE-METRICS.md`, `WEEKLY-PROGRESS-REPORT.md`) - Master QA Prompt for autonomous execution - README with complete quickstart guide **For autonomous execution** (recommended): See `references/master_qa_prompt.md` - single copy-paste command for 100x speedup. ## Core Capabilities ### 1. QA Project Initialization Initialize complete QA infrastructure with all templates: ```bash python scripts/init_qa_project.py <project-name> [output-directory] ``` Creates directory structure, tracking CSVs, documentation templates, and master prompt for autonomous execution. **Use when**: Starting QA from scratch or migrating to structured QA process. ### 2. Test Case Writing Write standardized, reproducible test cases following AAA pattern (Arrange-Act-Assert): 1. Read template: `assets/templates/TEST-CASE-TEMPLATE.md` 2. Follow structure: Prerequisites (Arrange) → Test Steps (Act) → Expected Results (Assert) 3. Assign priority: P0 (blocker) → P4 (low) 4. Include edge cases and potential bugs **Test case format**: TC-[CATEGORY]-[NUMBER] (e.g., TC-CLI-001, TC-WEB-042, TC-SEC-007) **Reference**: See `references/google_testing_standards.md` for complete AAA pattern guidelines and coverage thresholds. ### 3. Test Execution & Tracking **Ground Truth Principle** (critical): - **Test case documents** (e.g., `02-CLI-TEST-CASES.md`) = **authoritative source** for test steps - **Tracking CSV** = execution status only (do NOT trust CSV for test specifications) - See `references/ground_truth_principle.md` for preventing doc/CSV sync issues **Manual execution**: 1. Read test case from category document (e.g., `02-CLI-TEST-CASES.md`) ← **always start here** 2. Execute test steps exactly as documented 3. Update `TEST-EXECUTION-TRACKING.csv` **immediately** after EACH test (never batch) 4. File bug in `BUG-TRACKING-TEMPLATE.csv` if test fails **Autonomous execution** (recommended): 1. Copy master prompt from `references/master_qa_prompt.md` 2. Paste to LLM session 3. LLM auto-executes, auto-tracks, auto-files bugs, auto-generates reports **Innovation**: 100x faster vs manual + zero human error in tracking + auto-resume capability. ### 4. Bug Reporting File bugs with proper severity classification: **Required fields**: - Bug ID: Sequential (BUG-001, BUG-002, ...) - Severity: P0 (24h fix) → P4 (optional) - Steps to Reproduce: Numbered, specific - Environment: OS, versions, configuration **Severity classification**: - **P0 (Blocker)**: Security vulnerability, core functionality broken, data loss - **P1 (Critical)**: Major feature broken with workaround - **P2 (High)**: Minor feature issue, edge case - **P3 (Medium)**: Cosmetic issue - **P4 (Low)**: Documentation typo **Reference**: See `BUG-TRACKING-TEMPLATE.csv` for complete template with examples. ### 5. Quality Metrics Calculation Calculate comprehensive QA metrics and quality gates status: ```bash python scripts/calculate_metrics.py <path/to/TEST-EXECUTION-TRACKING.csv> ``` **Metrics dashboard includes**: - Test execution progress (X/Y tests, Z% complete) - Pass rate (passed/executed %) - Bug analysis (unique bugs, P0/P1/P2 breakdown) - Quality gates status (✅/❌ for each gate) **Quality gates** (all must pass for release): | Gate | Target | Blocker | |------|--------|---------| | Test Execution | 100% | Yes | | Pass Rate | ≥80% | Yes | | P0 Bugs | 0 | Yes | | P1 Bugs | ≤5 | Yes | | Code Coverage | ≥80% | Yes | | Security | 90% OWASP | Yes | ### 6. Progress Reporting Generate QA reports for stakeholders: **Daily summary** (end-of-day): - Tests executed, pass rate, bugs filed - Blockers (or None) - Tomorrow's plan **Weekly report** (every Friday): - Use template: `WEEKLY-PROGRESS-REPORT.md` (created by init script) - Compare against baseline: `BASELINE-METRICS.md` - Assess quality gates and trends **Reference**: See `references/llm_prompts_library.md` for 30+ ready-to-use reporting prompts. ### 7. Security Testing (OWASP) Implement OWASP Top 10 security testing: **Coverage targets**: 1. **A01: Broken Access Control** - RLS bypass, privilege escalation 2. **A02: Cryptographic Failures** - Token encryption, password hashing 3. **A03: Injection** - SQL injection, XSS, command injection 4. **A04: Insecure Design** - Rate limiting, anomaly detection 5. **A05: Security Misconfiguration** - Verbose errors, default credentials 6. **A07: Authentication Failures** - Session hijacking, CSRF 7. **Others**: Data integrity, logging, SSRF **Target**: 90% OWASP coverage (9/10 threats mitigated). Each security test follows AAA pattern with specific attack vectors documented. ## Day 1 Onboarding For new QA engineers joining a project, complete 5-hour onboarding guide: **Read**: `references/day1_onboarding.md` **Timeline**: - Hour 1: Environment setup (database, dev server, dependencies) - Hour 2: Documentation review (test strategy, quality gates) - Hour 3: Test data setup (users, CLI, DevTools) - Hour 4: Execute first test case - Hour 5: Team onboarding & Week 1 planning **Checkpoint**: By end of Day 1, environment running, first test executed, ready for Week 1. ## Autonomous Execution (⭐ Recommended) Enable LLM-driven autonomous QA testing with single master prompt: **Read**: `references/master_qa_prompt.md` **Features**: - Auto-resume from last completed test (reads tracking CSV) - Auto-execute test cases (Week 1-5 progression) - Auto-track results (updates CSV after each test) - Auto-file bugs (creates bug reports for failures) - Auto-generate reports (daily summaries, weekly reports) - Auto-escalate P0 bugs (stops testing, notifies stakeholders) **Benefits**: - 100x faster execution vs manual - Zero human error in tracking - Consistent bug documentation - Immediate progress visibility **Usage**: Copy master prompt, paste to LLM, let it run autonomously for 5 weeks. ## Adapting for Your Project ### Small Project (50 tests) - Timeline: 2 weeks - Categories: 2-3 (e.g., Frontend, Backend) - Daily: 5-7 tests - Reports: Daily summary only ### Medium Project (200 tests) - Timeline: 4 weeks - Categories: 4-5 (CLI, Web, API, DB, Security) - Daily: 10-12 tests - Reports: Daily + weekly ### Large Project (500+ tests) - Timeline: 8-10 weeks - Categories: 6-8 (multiple components) - Daily: 10-15 tests - Reports: Daily + weekly + bi-weekly stakeholder ## Reference Documents Access detailed guidelines from bundled references: - **`references/day1_onboarding.md`** - 5-hour onboarding guide for new QA engineers - **`references/master_qa_prompt.md`** - Single command for autonomous LLM execution (100x speedup) - **`references/llm_prompts_library.md`** - 30+ ready-to-use prompts for specific QA tasks - **`references/google_testing_standards.md`** - AAA pattern, coverage thresholds, fail-fast validation - **`references/ground_truth_principle.md`** - Preventing doc/CSV sync issues (critical for test suite integrity) ## Assets & Templates Test case templates and bug report formats: - **`assets/templates/TEST-CASE-TEMPLATE.md`** - Complete template with CLI and security examples ## Scripts Automation scripts for QA infrastructure: - **`scripts/init_qa_project.py`** - Initialize QA infrastructure (one command setup) - **`scripts/calculate_metrics.py`** - Generate quality metrics dashboard ## Common Patterns ### Pattern 1: Starting Fresh QA ``` 1. python scripts/init_qa_project.py my-app ./ 2. Fill in BASELINE-METRICS.md (document current state) 3. Write test cases using assets/templates/TEST-CASE-TEMPLATE.md 4. Copy master prompt from references/master_qa_prompt.md 5. Paste to LLM → autonomous execution begins ``` ### Pattern 2: LLM-Driven Testing (Autonomous) ``` 1. Read references/master_qa_prompt.md 2. Copy the single master prompt (one paragraph) 3. Paste to LLM conversation 4. LLM executes all 342 test cases over 5 weeks 5. LLM updates tracking CSVs automatically 6. LLM generates weekly reports automatically ``` ### Pattern 3: Adding Security Testing ``` 1. Read references/google_testing_standards.md (OWASP section) 2. Write TC-SEC-XXX test cases for each OWASP threat 3. Target 90% coverage (9/10 threats) 4. Document mitigations in test cases ``` ### Pattern 4: Third-Party QA Handoff ``` 1. Ensure all templates populated 2. Verify BASELINE-METRICS.md complete 3. Package tests/docs/ folder 4. Include references/master_qa_prompt.md for autonomous execution 5. QA team can start immediately (Day 1 onboarding → 5 weeks testing) ``` ## Success Criteria This skill is effective when: - ✅ Test cases are reproducible by any engineer - ✅ Quality gates objectively measured - ✅ Bugs fully documented with repro steps - ✅ Progress visible in real-time (CSV tracking) - ✅ Autonomous execution enabled (LLM can execute full plan) - ✅ Third-party QA teams can start testing immediately
Related Skills
vertex-infra-expert
Terraform infrastructure specialist for Vertex AI services and Gemini deployments. Provisions Model Garden, endpoints, vector search, pipelines, and enterprise AI infrastructure. Triggers: "vertex ai terraform", "gemini deployment terraform", "model garden infrastructure", "vertex ai endpoints"
validator-expert
Validate production readiness of Vertex AI Agent Engine deployments across security, monitoring, performance, compliance, and best practices. Generates weighted scores (0-100%) with actionable remediation plans. Use when asked to validate a deployment, run a production readiness check, audit security posture, or verify compliance for Vertex AI agents. Trigger with "validate deployment", "production readiness", "security audit", "compliance check", "is this agent ready for prod", "check my ADK agent", "review before deploy", or "production readiness check". Make sure to use this skill whenever validating ADK agents for Agent Engine.
genkit-production-expert
Build production Firebase Genkit applications including RAG systems, multi-step flows, and tool calling for Node.js/Python/Go. Deploy to Firebase Functions or Cloud Run with AI monitoring. Use when asked to "create genkit flow" or "implement RAG". Trigger with relevant phrases based on skill purpose.
genkit-infra-expert
Terraform infrastructure specialist for deploying Genkit applications to production. Provisions Firebase Functions, Cloud Run services, GKE clusters, monitoring, and CI/CD for Genkit AI workflows. Triggers: "deploy genkit terraform", "genkit infrastructure", "firebase functions terraform", "cloud run genkit"
gcp-examples-expert
Generate production-ready Google Cloud code examples from official repositories including ADK samples, Genkit templates, Vertex AI notebooks, and Gemini patterns. Use when asked to "show ADK example" or "provide GCP starter kit". Trigger with relevant phrases based on skill purpose.
adk-infra-expert
Terraform infrastructure specialist for Vertex AI ADK Agent Engine production deployments. Provisions Agent Engine runtime, Code Execution Sandbox, Memory Bank, VPC-SC, IAM, and secure multi-agent infrastructure. Triggers: "deploy adk terraform", "agent engine infrastructure", "adk production deployment", "vpc-sc agent engine"
paper-expert-generator
Generate a specialized domain-expert research agent modeled on PaperClaw architecture. Use this skill when a user wants to create an AI agent that can automatically search, filter, summarize, and evaluate academic papers in a specific research field. Trigger phrases include help me create a paper tracking agent for my field, I want an agent to monitor latest papers in bioinformatics, build me a paper review agent for computer vision, create a PaperClaw-style agent for my domain, generate a domain-specific paper expert agent. The generated agent is a complete OpenClaw agent with all required skills (arxiv-search, semantic-scholar, paper-review, daily-search, weekly-report) fully adapted for the target domain.
i18n-expert
This skill should be used when setting up, auditing, or enforcing internationalization/localization in UI codebases (React/TS, i18next or similar, JSON locales), including installing/configuring the i18n framework, replacing hard-coded strings, ensuring en-US/zh-CN coverage, mapping error codes to localized messages, and validating key parity, pluralization, and formatting.
pr-review-expert
Use when the user asks to review pull requests, analyze code changes, check for security issues in PRs, or assess code quality of diffs.
jira-expert
Atlassian Jira expert for creating and managing projects, planning, product discovery, JQL queries, workflows, custom fields, automation, reporting, and all Jira features. Use for Jira project setup, configuration, advanced search, dashboard creation, workflow design, and technical Jira operations.
../../../project-management/confluence-expert/SKILL.md
No description provided.
threat-modeling-expert
Expert in threat modeling methodologies, security architecture review, and risk assessment. Masters STRIDE, PASTA, attack trees, and security requirement extraction. Use for security architecture reviews, threat identification, and secure-by-design planning.