improving-python-coverage
Runs Python unit tests with coverage, analyzes coverage reports, and implements meaningful tests to increase coverage by ~0.2%. Use when you want to systematically improve Python test coverage with high-value test cases.
Best use case
improving-python-coverage is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
Runs Python unit tests with coverage, analyzes coverage reports, and implements meaningful tests to increase coverage by ~0.2%. Use when you want to systematically improve Python test coverage with high-value test cases.
Teams using improving-python-coverage should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/improving-python-coverage/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How improving-python-coverage Compares
| Feature / Agent | improving-python-coverage | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
Runs Python unit tests with coverage, analyzes coverage reports, and implements meaningful tests to increase coverage by ~0.2%. Use when you want to systematically improve Python test coverage with high-value test cases.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
Related Guides
AI Agents for Coding
Browse AI agent skills for coding, debugging, testing, refactoring, code review, and developer workflows across Claude, Cursor, and Codex.
AI Agents for Marketing
Discover AI agents for marketing workflows, from SEO and content production to campaign research, outreach, and analytics.
AI Agents for Startups
Explore AI agent skills for startup validation, product research, growth experiments, documentation, and fast execution with small teams.
SKILL.md Source
# Improving Python coverage Increase Python unit test coverage by ~0.2% through meaningful tests that add real value. **Be fully autonomous** — Do NOT stop or pause to ask for confirmation. Keep iterating (analyze → implement → verify) until the 0.2% coverage target is reached. If you encounter ambiguities about what to test, make a reasonable choice and proceed. ## Workflow **Step 1: Run tests with coverage** ```bash make python-tests # ~3 min, creates .coverage file ``` Generate JSON report for analysis: ```bash uv run coverage json -o coverage.json ``` The JSON contains per-file `missing_lines` arrays showing uncovered line numbers. **Step 2: Analyze and prioritize** Read `coverage.json` to find files with: 1. Large size + below-average `percent_covered` (high impact) 2. Core modules in `lib/streamlit/elements/` or `lib/streamlit/runtime/` 3. Pure utility functions Skip: >97% coverage, `proto/*`, `vendor/*`, `static/*`, test files. **Step 3: Implement tests (in subagent)** Launch a subagent to implement tests for each prioritized file. Provide the subagent with: - The target file path and its `missing_lines` from coverage - Instructions to read the source, existing tests, and write new tests - The test selection guidelines below The subagent should: 1. Read source and existing tests at `lib/tests/streamlit/<path>/<module>_test.py` 2. Write tests for: conditional branches, error handling, edge cases, exception paths 3. Follow `lib/tests/AGENTS.md`: prefer pytest-style standalone functions over `unittest.TestCase` classes, use `@pytest.mark.parametrize` to consolidate tests that only differ in inputs/expected outputs, add numpydoc docstrings and type annotations 4. Run the new tests to verify they pass: `uv run pytest lib/tests/streamlit/path/to/module_test.py -v` **Step 4: Verify and iterate** ```bash uv run pytest lib/tests/streamlit/path/to/module_test.py -v # Run new tests make python-tests # Measure progress ``` **Repeat steps 2-4 until coverage improves by ≥0.2%**, then run `make check`. **Step 5: Simplify, review, and address feedback** Once all tests pass and coverage target is met: 1. Run the `simplifying-local-changes` subagent to clean up and simplify the code changes. Wait for completion. 2. Run the `reviewing-local-changes` subagent to review the changes. Wait for completion and read the review output. 3. Address the review feedback: for each recommendation, implement it if valid and improves code quality; skip with brief reasoning if not applicable or would over-engineer. 4. Run /checking-changes to verify everything still passes after changes. ## Test selection **DO test:** Conditional logic, error handling, edge cases (None, empty, zero, max), public API functions, complex branches. **DON'T test:** Simple accessors, protobufs, implementation details, already well-covered code. **Coverage exclusions:** Use `# pragma: no cover` sparingly for code that genuinely doesn't need testing. Always include a reason (e.g., `# pragma: no cover - defensive`): - Platform-specific branches that can't run in CI (`# pragma: no cover - platform-specific`) - Defensive code that should never execute (`# pragma: no cover - defensive`) - Abstract method stubs or protocol definitions (`# pragma: no cover - abstract`) **Integration dependencies:** Packages listed under `[dependency-groups] integration` in `pyproject.toml` (e.g., `pydantic`, `sympy`, `polars`, `sqlalchemy`) are only installed for integration tests, not regular unit tests. When writing tests that use these packages: - Import them **inside the test function**, not at module top-level - Add `@pytest.mark.require_integration` marker to the test - This ensures tests gracefully skip when run outside the integration test environment ## Test file location `lib/tests/streamlit/<package>/<module>_test.py` mirrors `lib/streamlit/<package>/<module>.py` ## Notes - Quality > coverage numbers - skip tests that don't catch real bugs - Target is 95%+ coverage per `lib/tests/AGENTS.md` - Use `/checking-changes` after implementing tests - Some code paths involving external libraries (e.g., database connectors, optional dependencies) are already covered by integration tests marked with `pytest.mark.require_integration`. These integration tests are **not** included in the coverage numbers from `make python-tests`. When analyzing missing lines, check whether the uncovered code is exercised by integration tests before adding unit tests or `# pragma: no cover` annotations. - **Local vs CI coverage differences:** Code that is version-specific (Python version, library version) or uses integration dependencies may appear uncovered locally but is tested and covered in CI. Examples: - Python version-specific branches (e.g., `if sys.version_info >= (3, 14)`) run only on matching CI jobs - Library version-specific code (e.g., pandas 2.x vs 3.x behavior) is covered across CI matrix - Integration dependency tests (`@pytest.mark.require_integration`) run in separate CI jobs with those packages installed Before adding tests or `# pragma: no cover` for such code, verify whether it's already exercised in CI.
Related Skills
improving-frontend-coverage
Runs frontend unit tests with coverage, analyzes coverage reports, and implements meaningful tests to increase coverage by ~0.2%. Use when you want to systematically improve frontend test coverage with high-value test cases.
fixing-streamlit-ci
Analyze and fix failed GitHub Actions CI jobs for the current branch/PR. Use when CI checks fail, PR checks show failures, or you need to diagnose lint/type/test errors and verify fixes locally.
fixing-flaky-e2e-tests
Diagnose and fix flaky Playwright e2e tests. Use when tests fail intermittently, show timeout errors, have snapshot mismatches, or exhibit browser-specific failures.
finalizing-pr
Finalizes branch changes for merging by simplifying code, running checks, reviewing changes, and creating a PR if needed. Use when ready to merge changes into the target branch.
discovering-make-commands
Lists available make commands for Streamlit development. Use for build, test, lint, or format tasks.
debugging-streamlit
Debug Streamlit frontend and backend changes using make debug with hot-reload. Use when testing code changes, investigating bugs, checking UI behavior, or needing screenshots of the running app.
creating-pull-requests
Creates a draft pull request on GitHub with proper labels, branch naming, and description formatting. Use when changes are ready to be submitted as a PR to the streamlit/streamlit repository.
checking-changes
Validates all code changes before committing by running format, lint, type, and unit test checks. Use after making backend (Python) or frontend (TypeScript) changes, before committing or finishing a work session.
assessing-external-test-risk
Assesses whether branch or PR changes are high-risk for externally hosted or embedded Streamlit usage and recommends whether external e2e coverage with `@pytest.mark.external_test` is needed. Use during code review, PR triage, or test planning when changes touch routing, auth, websocket/session behavior, embedding, assets, cross-origin behavior, SiS/Snowflake runtime, storage, or security headers.
addressing-pr-review-comments
Address all valid review comments on a PR for the current branch in the streamlit/streamlit repo. Covers both inline review comments and general PR (issue) comments. Use when a PR has reviewer feedback to address, including code changes, style fixes, and documentation updates.
writing-spec
Writes product and tech specs for new Streamlit features. Use when designing new API commands, widgets, or significant changes that need team review before implementation.
updating-internal-docs
Review internal documentation (*.md files) against the current codebase state and propose updates for outdated or incorrect information.