Applicant Screening

## Overview

25 stars

Best use case

Applicant Screening is best used when you need a repeatable AI agent workflow instead of a one-off prompt.

## Overview

Teams using Applicant Screening should expect a more consistent output, faster repeated execution, less prompt rewriting.

When to use this skill

  • You want a reusable workflow that can be run more than once with consistent structure.

When not to use this skill

  • You only need a quick one-off answer and do not need a reusable workflow.
  • You cannot install or maintain the underlying files, dependencies, or repository context.

Installation

Claude Code / Cursor / Codex

$curl -o ~/.claude/skills/applicant-screening/SKILL.md --create-dirs "https://raw.githubusercontent.com/ComeOnOliver/skillshub/main/skills/TerminalSkills/skills/applicant-screening/SKILL.md"

Manual Installation

  1. Download SKILL.md from GitHub
  2. Place it in .claude/skills/applicant-screening/SKILL.md inside your project
  3. Restart your AI agent — it will auto-discover the skill

How Applicant Screening Compares

Feature / AgentApplicant ScreeningStandard Approach
Platform SupportNot specifiedLimited / Varies
Context Awareness High Baseline
Installation ComplexityUnknownN/A

Frequently Asked Questions

What does this skill do?

## Overview

Where can I find the source code?

You can find the source code on GitHub using the link provided at the top of the page.

SKILL.md Source

# Applicant Screening

## Overview

Screen job applications objectively by evaluating candidates against defined requirements. Build scoring rubrics from job descriptions, assess each candidate's qualifications, and produce ranked shortlists with clear justifications. Reduce bias by applying consistent criteria across all applicants.

## Instructions

When a user asks you to screen candidates or review applications, follow these steps:

### Step 1: Define the scoring rubric

Extract requirements from the job description and assign weights:

```yaml
rubric:
  role: Senior Backend Engineer
  total_points: 100

  required_criteria:
    - name: "Python experience (5+ years)"
      max_points: 20
      scoring:
        - { range: "7+ years", points: 20 }
        - { range: "5-7 years", points: 15 }
        - { range: "3-5 years", points: 8 }
        - { range: "<3 years", points: 0 }

    - name: "Distributed systems experience"
      max_points: 15
      scoring:
        - { range: "Led design of distributed systems", points: 15 }
        - { range: "Contributed to distributed systems", points: 10 }
        - { range: "Basic understanding", points: 5 }
        - { range: "No experience", points: 0 }

    - name: "Cloud platform experience (AWS/GCP/Azure)"
      max_points: 15
      scoring:
        - { range: "3+ years production experience", points: 15 }
        - { range: "1-3 years", points: 10 }
        - { range: "Certification only", points: 5 }
        - { range: "None", points: 0 }

  preferred_criteria:
    - name: "Team leadership/mentoring"
      max_points: 10
      scoring:
        - { range: "Managed team of 3+", points: 10 }
        - { range: "Mentored individuals", points: 6 }
        - { range: "None mentioned", points: 0 }

    - name: "System design skills"
      max_points: 10
      scoring:
        - { range: "Designed large-scale systems", points: 10 }
        - { range: "Some design experience", points: 5 }
        - { range: "None mentioned", points: 0 }

  education:
    - name: "Relevant degree"
      max_points: 10
      scoring:
        - { range: "MS/PhD in CS or related", points: 10 }
        - { range: "BS in CS or related", points: 7 }
        - { range: "Bootcamp or self-taught with strong portfolio", points: 5 }

  culture_fit:
    - name: "Communication quality"
      max_points: 10
      scoring:
        - { range: "Clear, well-structured application", points: 10 }
        - { range: "Adequate", points: 5 }
        - { range: "Poorly written", points: 2 }

    - name: "Role alignment"
      max_points: 10
      scoring:
        - { range: "Clear interest in this specific role", points: 10 }
        - { range: "Generic application", points: 4 }
```

Present the rubric to the user for approval before screening.

### Step 2: Screen each candidate

For each application, evaluate against every criterion:

```
Candidate: Alice Chen
Resume: alice_chen_resume.pdf

Evaluation:
  Python experience: 20/20 - 8 years of Python at two companies
  Distributed systems: 15/15 - Led redesign of event-driven architecture
  Cloud platform: 10/15 - 2 years AWS, no multi-cloud experience
  Team leadership: 10/10 - Managed team of 5 engineers
  System design: 10/10 - Designed payment processing system at scale
  Relevant degree: 7/10 - BS Computer Science, Stanford
  Communication: 10/10 - Well-structured resume, clear achievements
  Role alignment: 8/10 - Cover letter references specific team projects

  TOTAL: 90/100
  Recommendation: STRONG YES - Advance to interview
```

### Step 3: Generate the ranked shortlist

```
SCREENING RESULTS - Senior Backend Engineer
============================================
Screened: 15 candidates
Date: 2025-01-15

SHORTLIST (Score >= 70):
  1. Alice Chen        - 90/100 - STRONG YES
  2. Marcus Johnson    - 85/100 - STRONG YES
  3. Priya Patel       - 78/100 - YES
  4. David Kim         - 72/100 - YES

MAYBE (Score 50-69):
  5. Sarah Williams    - 65/100 - Lacks distributed systems exp
  6. Tom Brown         - 58/100 - Junior for role level

DECLINE (Score < 50):
  7-15. [8 candidates below threshold]

NOTES:
  - Top 4 candidates meet all required criteria
  - Alice Chen and Marcus Johnson are standout candidates
  - Consider Sarah Williams if pipeline needs expansion
```

### Step 4: Save results

Save the full screening report:

```bash
# Save detailed report
cat > screening_report.md << 'EOF'
[full report with individual evaluations]
EOF

# Save summary CSV for tracking
cat > screening_summary.csv << 'EOF'
candidate,score,recommendation,top_strength,gap
Alice Chen,90,Strong Yes,Distributed systems,None
Marcus Johnson,85,Strong Yes,Python expertise,Limited cloud
EOF
```

## Examples

### Example 1: Screen a batch of resumes

**User request:** "I have 20 resumes for our frontend developer role. Help me create a shortlist."

**Steps:**
1. Read the job description to build the rubric
2. Present the rubric for user approval
3. Read each resume file
4. Score each candidate against the rubric
5. Generate ranked shortlist with top 5-7 candidates
6. Save the report as `screening_report.md`

### Example 2: Create a custom scoring rubric

**User request:** "Build me a screening rubric for a product manager role that weighs user research experience heavily."

**Output:**
```yaml
rubric:
  role: Product Manager
  total_points: 100
  required_criteria:
    - name: "User research experience"
      max_points: 25  # Heavily weighted per request
    - name: "Product lifecycle management"
      max_points: 20
    - name: "Data-driven decision making"
      max_points: 15
    - name: "Stakeholder management"
      max_points: 15
  preferred_criteria:
    - name: "Technical background"
      max_points: 10
    - name: "Industry experience"
      max_points: 10
    - name: "Communication quality"
      max_points: 5
```

### Example 3: Re-evaluate with adjusted criteria

**User request:** "We decided Kubernetes experience is now required. Re-screen the candidates."

**Steps:**
1. Add Kubernetes as a required criterion (15 points)
2. Rebalance other criteria to maintain 100-point total
3. Re-evaluate all candidates against updated rubric
4. Generate updated shortlist and highlight ranking changes

## Guidelines

- Always present the scoring rubric to the user for approval before screening candidates.
- Apply the same rubric consistently to every candidate. Do not adjust mid-screening.
- Base scores only on evidence present in the application. Do not infer or assume qualifications.
- Note when a candidate's experience is ambiguous and flag it for the hiring manager to clarify.
- Remove identifying information (name, gender, age, photo) from the evaluation if the user requests blind screening.
- Never make final hiring decisions. Present scored recommendations and let the hiring team decide.
- Flag potential biases: if all top candidates share a background, note this for the user's awareness.
- Keep individual evaluations factual. Use "Resume does not mention X" rather than "Candidate lacks X."
- Save all screening artifacts so the process is auditable and repeatable.
- For large batches (20+ candidates), do a quick pre-screen first to eliminate clearly unqualified applicants before detailed scoring.

Related Skills

Daily Logs

25
from ComeOnOliver/skillshub

Record the user's daily activities, progress, decisions, and learnings in a structured, chronological format.

Socratic Method: The Dialectic Engine

25
from ComeOnOliver/skillshub

This skill transforms Claude into a Socratic agent — a cognitive partner who guides

Sokratische Methode: Die Dialektik-Maschine

25
from ComeOnOliver/skillshub

Dieser Skill verwandelt Claude in einen sokratischen Agenten — einen kognitiven Partner, der Nutzende durch systematisches Fragen zur Wissensentdeckung führt, anstatt direkt zu instruieren.

College Football Data (CFB)

25
from ComeOnOliver/skillshub

Before writing queries, consult `references/api-reference.md` for endpoints, conference IDs, team IDs, and data shapes.

College Basketball Data (CBB)

25
from ComeOnOliver/skillshub

Before writing queries, consult `references/api-reference.md` for endpoints, conference IDs, team IDs, and data shapes.

Betting Analysis

25
from ComeOnOliver/skillshub

Before writing queries, consult `references/api-reference.md` for odds formats, command parameters, and key concepts.

Research Proposal Generator

25
from ComeOnOliver/skillshub

Generate high-quality academic research proposals for PhD applications following Nature Reviews-style academic writing conventions.

Paper Slide Deck Generator

25
from ComeOnOliver/skillshub

Transform academic papers and content into professional slide deck images with automatic figure extraction.

Medical Imaging AI Literature Review Skill

25
from ComeOnOliver/skillshub

Write comprehensive literature reviews following a systematic 7-phase workflow.

Meeting Briefing Skill

25
from ComeOnOliver/skillshub

You are a meeting preparation assistant for an in-house legal team. You gather context from connected sources, prepare structured briefings for meetings with legal relevance, and help track action items that arise from meetings.

Canned Responses Skill

25
from ComeOnOliver/skillshub

You are a response template assistant for an in-house legal team. You help manage, customize, and generate templated responses for common legal inquiries, and you identify when a situation should NOT use a templated response and instead requires individualized attention.

Copywriting

25
from ComeOnOliver/skillshub

## Purpose