senior-prompt-engineer
World-class prompt engineering skill for LLM optimization, prompt patterns, structured outputs, and AI product development. Expertise in Claude, GPT-4, prompt design patterns, few-shot learning, chain-of-thought, and AI evaluation. Includes RAG optimization, agent design, and LLM system architecture. Use when building AI products, optimizing LLM performance, designing agentic systems, or implementing advanced prompting techniques.
Best use case
senior-prompt-engineer is best used when you need a repeatable AI agent workflow instead of a one-off prompt.
World-class prompt engineering skill for LLM optimization, prompt patterns, structured outputs, and AI product development. Expertise in Claude, GPT-4, prompt design patterns, few-shot learning, chain-of-thought, and AI evaluation. Includes RAG optimization, agent design, and LLM system architecture. Use when building AI products, optimizing LLM performance, designing agentic systems, or implementing advanced prompting techniques.
Teams using senior-prompt-engineer should expect a more consistent output, faster repeated execution, less prompt rewriting.
When to use this skill
- You want a reusable workflow that can be run more than once with consistent structure.
When not to use this skill
- You only need a quick one-off answer and do not need a reusable workflow.
- You cannot install or maintain the underlying files, dependencies, or repository context.
Installation
Claude Code / Cursor / Codex
Manual Installation
- Download SKILL.md from GitHub
- Place it in
.claude/skills/senior-prompt-engineer/SKILL.mdinside your project - Restart your AI agent — it will auto-discover the skill
How senior-prompt-engineer Compares
| Feature / Agent | senior-prompt-engineer | Standard Approach |
|---|---|---|
| Platform Support | Not specified | Limited / Varies |
| Context Awareness | High | Baseline |
| Installation Complexity | Unknown | N/A |
Frequently Asked Questions
What does this skill do?
World-class prompt engineering skill for LLM optimization, prompt patterns, structured outputs, and AI product development. Expertise in Claude, GPT-4, prompt design patterns, few-shot learning, chain-of-thought, and AI evaluation. Includes RAG optimization, agent design, and LLM system architecture. Use when building AI products, optimizing LLM performance, designing agentic systems, or implementing advanced prompting techniques.
Where can I find the source code?
You can find the source code on GitHub using the link provided at the top of the page.
Related Guides
SKILL.md Source
# Senior Prompt Engineer World-class senior prompt engineer skill for production-grade AI/ML/Data systems. ## Quick Start ### Main Capabilities ```bash # Core Tool 1 python scripts/prompt_optimizer.py --input data/ --output results/ # Core Tool 2 python scripts/rag_evaluator.py --target project/ --analyze # Core Tool 3 python scripts/agent_orchestrator.py --config config.yaml --deploy ``` ## Core Expertise This skill covers world-class capabilities in: - Advanced production patterns and architectures - Scalable system design and implementation - Performance optimization at scale - MLOps and DataOps best practices - Real-time processing and inference - Distributed computing frameworks - Model deployment and monitoring - Security and compliance - Cost optimization - Team leadership and mentoring ## Tech Stack **Languages:** Python, SQL, R, Scala, Go **ML Frameworks:** PyTorch, TensorFlow, Scikit-learn, XGBoost **Data Tools:** Spark, Airflow, dbt, Kafka, Databricks **LLM Frameworks:** LangChain, LlamaIndex, DSPy **Deployment:** Docker, Kubernetes, AWS/GCP/Azure **Monitoring:** MLflow, Weights & Biases, Prometheus **Databases:** PostgreSQL, BigQuery, Snowflake, Pinecone ## Reference Documentation ### 1. Prompt Engineering Patterns Comprehensive guide available in `references/prompt_engineering_patterns.md` covering: - Advanced patterns and best practices - Production implementation strategies - Performance optimization techniques - Scalability considerations - Security and compliance - Real-world case studies ### 2. Llm Evaluation Frameworks Complete workflow documentation in `references/llm_evaluation_frameworks.md` including: - Step-by-step processes - Architecture design patterns - Tool integration guides - Performance tuning strategies - Troubleshooting procedures ### 3. Agentic System Design Technical reference guide in `references/agentic_system_design.md` with: - System design principles - Implementation examples - Configuration best practices - Deployment strategies - Monitoring and observability ## Production Patterns ### Pattern 1: Scalable Data Processing Enterprise-scale data processing with distributed computing: - Horizontal scaling architecture - Fault-tolerant design - Real-time and batch processing - Data quality validation - Performance monitoring ### Pattern 2: ML Model Deployment Production ML system with high availability: - Model serving with low latency - A/B testing infrastructure - Feature store integration - Model monitoring and drift detection - Automated retraining pipelines ### Pattern 3: Real-Time Inference High-throughput inference system: - Batching and caching strategies - Load balancing - Auto-scaling - Latency optimization - Cost optimization ## Best Practices ### Development - Test-driven development - Code reviews and pair programming - Documentation as code - Version control everything - Continuous integration ### Production - Monitor everything critical - Automate deployments - Feature flags for releases - Canary deployments - Comprehensive logging ### Team Leadership - Mentor junior engineers - Drive technical decisions - Establish coding standards - Foster learning culture - Cross-functional collaboration ## Performance Targets **Latency:** - P50: < 50ms - P95: < 100ms - P99: < 200ms **Throughput:** - Requests/second: > 1000 - Concurrent users: > 10,000 **Availability:** - Uptime: 99.9% - Error rate: < 0.1% ## Security & Compliance - Authentication & authorization - Data encryption (at rest & in transit) - PII handling and anonymization - GDPR/CCPA compliance - Regular security audits - Vulnerability management ## Common Commands ```bash # Development python -m pytest tests/ -v --cov python -m black src/ python -m pylint src/ # Training python scripts/train.py --config prod.yaml python scripts/evaluate.py --model best.pth # Deployment docker build -t service:v1 . kubectl apply -f k8s/ helm upgrade service ./charts/ # Monitoring kubectl logs -f deployment/service python scripts/health_check.py ``` ## Resources - Advanced Patterns: `references/prompt_engineering_patterns.md` - Implementation Guide: `references/llm_evaluation_frameworks.md` - Technical Reference: `references/agentic_system_design.md` - Automation Scripts: `scripts/` directory ## Senior-Level Responsibilities As a world-class senior professional: 1. **Technical Leadership** - Drive architectural decisions - Mentor team members - Establish best practices - Ensure code quality 2. **Strategic Thinking** - Align with business goals - Evaluate trade-offs - Plan for scale - Manage technical debt 3. **Collaboration** - Work across teams - Communicate effectively - Build consensus - Share knowledge 4. **Innovation** - Stay current with research - Experiment with new approaches - Contribute to community - Drive continuous improvement 5. **Production Excellence** - Ensure high availability - Monitor proactively - Optimize performance - Respond to incidents
Related Skills
conducting-chaos-engineering
This skill enables Claude to design and execute chaos engineering experiments to test system resilience. It is used when the user requests help with failure injection, latency simulation, resource exhaustion testing, or resilience validation. The skill is triggered by discussions of chaos experiments (GameDays), failure injection strategies, resilience testing, and validation of recovery mechanisms like circuit breakers and retry logic. It leverages tools like Chaos Mesh, Gremlin, Toxiproxy, and AWS FIS to simulate real-world failures and assess system behavior.
optimizing-prompts
This skill optimizes prompts for Large Language Models (LLMs) to reduce token usage, lower costs, and improve performance. It analyzes the prompt, identifies areas for simplification and redundancy removal, and rewrites the prompt to be more concise and effective. It is used when the user wants to reduce LLM costs, improve response speed, or enhance the quality of LLM outputs by optimizing the prompt. Trigger terms include "optimize prompt", "reduce LLM cost", "improve prompt performance", "rewrite prompt", "prompt optimization".
adk-engineer
Execute software engineer specializing in creating production-ready ADK agents with best practices, code structure, testing, and deployment automation. Use when asked to "build ADK agent", "create agent code", or "engineer ADK application". Trigger with relevant phrases based on skill purpose.
promptify
Transform user requests into detailed, precise prompts for AI models. Use when users say "promptify", "promptify this", or explicitly request prompt engineering or improvement of their request for better AI responses.
prompt-improver
Optimize prompts for better AI responses. Use when user asks to improve a prompt, refine a prompt, make a prompt better, optimize prompting, review their prompt, or says "/improve-prompt". Transforms vague requests into clear, specific, actionable prompts.
gws-modelarmor-sanitize-prompt
Google Model Armor: Sanitize a user prompt through a Model Armor template.
tldr-prompt
Create tldr summaries for GitHub Copilot files (prompts, agents, instructions, collections), MCP servers, or documentation from URLs and queries.
prompt-builder
Guide users through creating high-quality GitHub Copilot prompts with proper structure, tools, and best practices.
ROS 2 Engineering Skills
A progressive-disclosure skill for ROS 2 development — from first workspace to
using-dbt-for-analytics-engineering
Builds and modifies dbt models, writes SQL transformations using ref() and source(), creates tests, and validates results with dbt show. Use when doing any dbt work - building or modifying models, debugging errors, exploring unfamiliar data sources, writing tests, or evaluating impact of changes.
promptfoo-evaluation
Configures and runs LLM evaluation using Promptfoo framework. Use when setting up prompt testing, creating evaluation configs (promptfooconfig.yaml), writing Python custom assertions, implementing llm-rubric for LLM-as-judge, or managing few-shot examples in prompts. Triggers on keywords like "promptfoo", "eval", "LLM evaluation", "prompt testing", or "model comparison".
prompt-injection-test
A test skill with prompt injection patterns