Best use case
Applicant Screening is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
## Overview
Teams using Applicant Screening should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/applicant-screening/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How Applicant Screening Compares
| Feature / Agent | Applicant Screening | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
## Overview
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
SKILL.md Source
# Applicant Screening
## Overview
Screen job applications objectively by evaluating candidates against defined requirements. Build scoring rubrics from job descriptions, assess each candidate's qualifications, and produce ranked shortlists with clear justifications. Reduce bias by applying consistent criteria across all applicants.
## Instructions
When a user asks you to screen candidates or review applications, follow these steps:
### Step 1: Define the scoring rubric
Extract requirements from the job description and assign weights:
```yaml
rubric:
role: Senior Backend Engineer
total_points: 100
required_criteria:
- name: "Python experience (5+ years)"
max_points: 20
scoring:
- { range: "7+ years", points: 20 }
- { range: "5-7 years", points: 15 }
- { range: "3-5 years", points: 8 }
- { range: "<3 years", points: 0 }
- name: "Distributed systems experience"
max_points: 15
scoring:
- { range: "Led design of distributed systems", points: 15 }
- { range: "Contributed to distributed systems", points: 10 }
- { range: "Basic understanding", points: 5 }
- { range: "No experience", points: 0 }
- name: "Cloud platform experience (AWS/GCP/Azure)"
max_points: 15
scoring:
- { range: "3+ years production experience", points: 15 }
- { range: "1-3 years", points: 10 }
- { range: "Certification only", points: 5 }
- { range: "None", points: 0 }
preferred_criteria:
- name: "Team leadership/mentoring"
max_points: 10
scoring:
- { range: "Managed team of 3+", points: 10 }
- { range: "Mentored individuals", points: 6 }
- { range: "None mentioned", points: 0 }
- name: "System design skills"
max_points: 10
scoring:
- { range: "Designed large-scale systems", points: 10 }
- { range: "Some design experience", points: 5 }
- { range: "None mentioned", points: 0 }
education:
- name: "Relevant degree"
max_points: 10
scoring:
- { range: "MS/PhD in CS or related", points: 10 }
- { range: "BS in CS or related", points: 7 }
- { range: "Bootcamp or self-taught with strong portfolio", points: 5 }
culture_fit:
- name: "Communication quality"
max_points: 10
scoring:
- { range: "Clear, well-structured application", points: 10 }
- { range: "Adequate", points: 5 }
- { range: "Poorly written", points: 2 }
- name: "Role alignment"
max_points: 10
scoring:
- { range: "Clear interest in this specific role", points: 10 }
- { range: "Generic application", points: 4 }
```
Present the rubric to the user for approval before screening.
### Step 2: Screen each candidate
For each application, evaluate against every criterion:
```
Candidate: Alice Chen
Resume: alice_chen_resume.pdf
Evaluation:
Python experience: 20/20 - 8 years of Python at two companies
Distributed systems: 15/15 - Led redesign of event-driven architecture
Cloud platform: 10/15 - 2 years AWS, no multi-cloud experience
Team leadership: 10/10 - Managed team of 5 engineers
System design: 10/10 - Designed payment processing system at scale
Relevant degree: 7/10 - BS Computer Science, Stanford
Communication: 10/10 - Well-structured resume, clear achievements
Role alignment: 8/10 - Cover letter references specific team projects
TOTAL: 90/100
Recommendation: STRONG YES - Advance to interview
```
### Step 3: Generate the ranked shortlist
```
SCREENING RESULTS - Senior Backend Engineer
============================================
Screened: 15 candidates
Date: 2025-01-15
SHORTLIST (Score >= 70):
1. Alice Chen - 90/100 - STRONG YES
2. Marcus Johnson - 85/100 - STRONG YES
3. Priya Patel - 78/100 - YES
4. David Kim - 72/100 - YES
MAYBE (Score 50-69):
5. Sarah Williams - 65/100 - Lacks distributed systems exp
6. Tom Brown - 58/100 - Junior for role level
DECLINE (Score < 50):
7-15. [8 candidates below threshold]
NOTES:
- Top 4 candidates meet all required criteria
- Alice Chen and Marcus Johnson are standout candidates
- Consider Sarah Williams if pipeline needs expansion
```
### Step 4: Save results
Save the full screening report:
```bash
# Save detailed report
cat > screening_report.md << 'EOF'
[full report with individual evaluations]
EOF
# Save summary CSV for tracking
cat > screening_summary.csv << 'EOF'
candidate,score,recommendation,top_strength,gap
Alice Chen,90,Strong Yes,Distributed systems,None
Marcus Johnson,85,Strong Yes,Python expertise,Limited cloud
EOF
```
## Examples
### Example 1: Screen a batch of resumes
**User request:** "I have 20 resumes for our frontend developer role. Help me create a shortlist."
**Steps:**
1. Read the job description to build the rubric
2. Present the rubric for user approval
3. Read each resume file
4. Score each candidate against the rubric
5. Generate ranked shortlist with top 5-7 candidates
6. Save the report as `screening_report.md`
### Example 2: Create a custom scoring rubric
**User request:** "Build me a screening rubric for a product manager role that weighs user research experience heavily."
**Output:**
```yaml
rubric:
role: Product Manager
total_points: 100
required_criteria:
- name: "User research experience"
max_points: 25 # Heavily weighted per request
- name: "Product lifecycle management"
max_points: 20
- name: "Data-driven decision making"
max_points: 15
- name: "Stakeholder management"
max_points: 15
preferred_criteria:
- name: "Technical background"
max_points: 10
- name: "Industry experience"
max_points: 10
- name: "Communication quality"
max_points: 5
```
### Example 3: Re-evaluate with adjusted criteria
**User request:** "We decided Kubernetes experience is now required. Re-screen the candidates."
**Steps:**
1. Add Kubernetes as a required criterion (15 points)
2. Rebalance other criteria to maintain 100-point total
3. Re-evaluate all candidates against updated rubric
4. Generate updated shortlist and highlight ranking changes
## Guidelines
- Always present the scoring rubric to the user for approval before screening candidates.
- Apply the same rubric consistently to every candidate. Do not adjust mid-screening.
- Base scores only on evidence present in the application. Do not infer or assume qualifications.
- Note when a candidate's experience is ambiguous and flag it for the hiring manager to clarify.
- Remove identifying information (name, gender, age, photo) from the evaluation if the user requests blind screening.
- Never make final hiring decisions. Present scored recommendations and let the hiring team decide.
- Flag potential biases: if all top candidates share a background, note this for the user's awareness.
- Keep individual evaluations factual. Use "Resume does not mention X" rather than "Candidate lacks X."
- Save all screening artifacts so the process is auditable and repeatable.
- For large batches (20+ candidates), do a quick pre-screen first to eliminate clearly unqualified applicants before detailed scoring.Related Skills
Daily Logs
Record the user's daily activities, progress, decisions, and learnings in a structured, chronological format.
Socratic Method: The Dialectic Engine
This skill transforms Claude into a Socratic agent — a cognitive partner who guides
Sokratische Methode: Die Dialektik-Maschine
Dieser Skill verwandelt Claude in einen sokratischen Agenten — einen kognitiven Partner, der Nutzende durch systematisches Fragen zur Wissensentdeckung führt, anstatt direkt zu instruieren.
College Football Data (CFB)
Before writing queries, consult `references/api-reference.md` for endpoints, conference IDs, team IDs, and data shapes.
College Basketball Data (CBB)
Before writing queries, consult `references/api-reference.md` for endpoints, conference IDs, team IDs, and data shapes.
Betting Analysis
Before writing queries, consult `references/api-reference.md` for odds formats, command parameters, and key concepts.
Research Proposal Generator
Generate high-quality academic research proposals for PhD applications following Nature Reviews-style academic writing conventions.
Paper Slide Deck Generator
Transform academic papers and content into professional slide deck images with automatic figure extraction.
Medical Imaging AI Literature Review Skill
Write comprehensive literature reviews following a systematic 7-phase workflow.
Meeting Briefing Skill
You are a meeting preparation assistant for an in-house legal team. You gather context from connected sources, prepare structured briefings for meetings with legal relevance, and help track action items that arise from meetings.
Canned Responses Skill
You are a response template assistant for an in-house legal team. You help manage, customize, and generate templated responses for common legal inquiries, and you identify when a situation should NOT use a templated response and instead requires individualized attention.
Copywriting
## Purpose