提示工程
prompt-engineering
by giuseppe-trisciuoglio
聚焦高级 Prompt 设计与优化,覆盖 few-shot、chain-of-thought、模板体系和 system prompt,适合为复杂推理任务构建可复用、可评估、可上线的生产级提示方案。
安装
claude skill add --url github.com/giuseppe-trisciuoglio/developer-kit/tree/main/plugins/developer-kit-ai/skills/prompt-engineering文档
Prompt Engineering
Overview
This skill provides comprehensive frameworks for creating, optimizing, and implementing advanced prompt patterns that significantly improve LLM performance across various tasks and models. It covers few-shot learning, chain-of-thought reasoning, prompt optimization workflows, template systems, and system prompt design for production-ready AI applications.
When to Use
Use this skill when:
- Creating new prompts for complex reasoning or analytical tasks
- Optimizing existing prompts for better accuracy or efficiency
- Implementing few-shot learning with strategic example selection
- Designing chain-of-thought reasoning for multi-step problems
- Building reusable prompt templates and systems
- Developing system prompts for consistent model behavior
- Troubleshooting poor prompt performance or failure modes
- Scaling prompt systems for production use cases
Core Prompt Engineering Patterns
1. Few-Shot Learning Implementation
Select examples using semantic similarity and diversity sampling to maximize learning within context window constraints.
Example Selection Strategy
- Use
references/few-shot-patterns.mdfor comprehensive selection frameworks - Balance example count (3-5 optimal) with context window limitations
- Include edge cases and boundary conditions in example sets
- Prioritize diverse examples that cover problem space variations
- Order examples from simple to complex for progressive learning
Few-Shot Template Structure
Example 1 (Basic case):
Input: {representative_input}
Output: {expected_output}
Example 2 (Edge case):
Input: {challenging_input}
Output: {robust_output}
Example 3 (Error case):
Input: {problematic_input}
Output: {corrected_output}
Now handle: {target_input}
2. Chain-of-Thought Reasoning
Elicit step-by-step reasoning for complex problem-solving through structured thinking patterns.
Implementation Patterns
- Reference
references/cot-patterns.mdfor detailed reasoning frameworks - Use "Let's think step by step" for zero-shot CoT initiation
- Provide complete reasoning traces for few-shot CoT demonstrations
- Implement self-consistency by sampling multiple reasoning paths
- Include verification and validation steps in reasoning chains
CoT Template Structure
Let's approach this step-by-step:
Step 1: {break_down_the_problem}
Analysis: {detailed_reasoning}
Step 2: {identify_key_components}
Analysis: {component_analysis}
Step 3: {synthesize_solution}
Analysis: {solution_justification}
Final Answer: {conclusion_with_confidence}
3. Prompt Optimization Workflows
Implement iterative refinement processes with measurable performance metrics and systematic A/B testing.
Optimization Process
- Use
references/optimization-frameworks.mdfor comprehensive optimization strategies - Measure baseline performance before optimization attempts
- Implement single-variable changes for accurate attribution
- Track metrics: accuracy, consistency, latency, token efficiency
- Use statistical significance testing for A/B validation
- Document optimization iterations and their impacts
Performance Metrics Framework
- Accuracy: Task completion rate and output correctness
- Consistency: Response stability across multiple runs
- Efficiency: Token usage and response time optimization
- Robustness: Performance across edge cases and variations
- Safety: Adherence to guidelines and harm prevention
4. Template Systems Architecture
Build modular, reusable prompt components with variable interpolation and conditional sections.
Template Design Principles
- Reference
references/template-systems.mdfor modular template frameworks - Use clear variable naming conventions (e.g.,
{user_input},{context}) - Implement conditional sections for different scenario handling
- Design role-based templates for specific use cases
- Create hierarchical template composition patterns
Template Structure Example
# System Context
You are a {role} with {expertise_level} expertise in {domain}.
# Task Context
{if background_information}
Background: {background_information}
{endif}
# Instructions
{task_instructions}
# Examples
{example_count}
# Output Format
{output_specification}
# Input
{user_query}
5. System Prompt Design
Design comprehensive system prompts that establish consistent model behavior, output formats, and safety constraints.
System Prompt Components
- Use
references/system-prompt-design.mdfor detailed design guidelines - Define clear role specification and expertise boundaries
- Establish output format requirements and structural constraints
- Include safety guidelines and content policy adherence
- Set context for background information and domain knowledge
System Prompt Framework
You are an expert {role} specializing in {domain} with {experience_level} of experience.
## Core Capabilities
- List specific capabilities and expertise areas
- Define scope of knowledge and limitations
## Behavioral Guidelines
- Specify interaction style and communication approach
- Define error handling and uncertainty protocols
- Establish quality standards and verification requirements
## Output Requirements
- Specify format expectations and structural requirements
- Define content inclusion and exclusion criteria
- Establish consistency and validation requirements
## Safety and Ethics
- Include content policy adherence
- Specify bias mitigation requirements
- Define harm prevention protocols
Implementation Workflows
Workflow 1: Create New Prompt from Requirements
-
Analyze Requirements
- Identify task complexity and reasoning requirements
- Determine target model capabilities and limitations
- Define success criteria and evaluation metrics
- Assess need for few-shot learning or CoT reasoning
-
Select Pattern Strategy
- Use few-shot learning for classification or transformation tasks
- Apply CoT for complex reasoning or multi-step problems
- Implement template systems for reusable prompt architecture
- Design system prompts for consistent behavior requirements
-
Draft Initial Prompt
- Structure prompt with clear sections and logical flow
- Include relevant examples or reasoning demonstrations
- Specify output format and quality requirements
- Incorporate safety guidelines and constraints
-
Validate and Test
- Test with diverse input scenarios including edge cases
- Measure performance against defined success criteria
- Iterate refinement based on testing results
- Document optimization decisions and their rationale
Workflow 2: Optimize Existing Prompt
-
Performance Analysis
- Measure current prompt performance metrics
- Identify failure modes and error patterns
- Analyze token efficiency and response latency
- Assess consistency across multiple runs
-
Optimization Strategy
- Apply systematic A/B testing with single-variable changes
- Use few-shot learning to improve task adherence
- Implement CoT reasoning for complex task components
- Refine template structure for better clarity
-
Implementation and Testing
- Deploy optimized prompts with controlled rollout
- Monitor performance metrics in production environment
- Compare against baseline using statistical significance
- Document improvements and lessons learned
Workflow 3: Scale Prompt Systems
-
Modular Architecture Design
- Decompose complex prompts into reusable components
- Create template inheritance hierarchies
- Implement dynamic example selection systems
- Build automated quality assurance frameworks
-
Production Integration
- Implement prompt versioning and rollback capabilities
- Create performance monitoring and alerting systems
- Build automated testing frameworks for prompt validation
- Establish update and deployment workflows
Quality Assurance
Validation Requirements
- Test prompts with at least 10 diverse scenarios
- Include edge cases, boundary conditions, and failure modes
- Verify output format compliance and structural consistency
- Validate safety guideline adherence and harm prevention
- Measure performance across multiple model runs
Performance Standards
- Achieve >90% task completion for well-defined use cases
- Maintain <5% variance across multiple runs for consistency
- Optimize token usage without sacrificing accuracy
- Ensure response latency meets application requirements
- Demonstrate robust handling of edge cases and unexpected inputs
Integration with Other Skills
This skill integrates seamlessly with:
- langchain4j-ai-services-patterns: Interface-based prompt design
- langchain4j-rag-implementation-patterns: Context-enhanced prompting
- langchain4j-testing-strategies: Prompt validation frameworks
- unit-test-parameterized: Systematic prompt testing approaches
Resources and References
references/few-shot-patterns.md: Comprehensive few-shot learning frameworksreferences/cot-patterns.md: Chain-of-thought reasoning patterns and examplesreferences/optimization-frameworks.md: Systematic prompt optimization methodologiesreferences/template-systems.md: Modular template design and implementationreferences/system-prompt-design.md: System prompt architecture and best practices
Usage Examples
Example 1: Classification Task with Few-Shot Learning
Classify customer feedback into categories using semantic similarity for example selection and diversity sampling for edge case coverage.
Example 2: Complex Reasoning with Chain-of-Thought
Implement step-by-step reasoning for financial analysis with verification steps and confidence scoring.
Example 3: Template System for Customer Service
Create modular templates with role-based components and conditional sections for different inquiry types.
Example 4: System Prompt for Code Generation
Design comprehensive system prompt with behavioral guidelines, output requirements, and safety constraints.
Common Pitfalls and Solutions
- Overfitting examples: Use diverse example sets with semantic variety
- Context window overflow: Implement strategic example selection and compression
- Inconsistent outputs: Specify clear output formats and validation requirements
- Poor generalization: Include edge cases and boundary conditions in training examples
- Safety violations: Incorporate comprehensive content policies and harm prevention
Performance Optimization
- Monitor token usage and implement compression strategies
- Use caching for repeated prompt components
- Optimize example selection for maximum learning efficiency
- Implement progressive disclosure for complex prompt systems
- Balance prompt complexity with response quality requirements
Best Practices
Design Principles
- Write clear, specific instructions that leave no ambiguity
- Structure prompts with logical sections and consistent formatting
- Use role-based prompts for consistent behavioral patterns
- Include concrete examples to demonstrate expected outputs
- Define output format requirements explicitly
Implementation Guidelines
- Start with simple prompts and iterate based on testing
- Test with diverse input scenarios including edge cases
- Document prompt versions and their performance metrics
- Implement error handling for unexpected inputs
- Monitor performance in production environments
Common Pitfalls to Avoid
- Overfitting examples to specific patterns
- Ignoring context window limitations and token budgeting
- Neglecting output format specifications
- Failing to handle edge cases and error scenarios
- Not measuring and tracking prompt performance over time
Constraints and Warnings
Model Limitations
- Different models have varying capabilities and token limits
- Complex prompts may exceed context windows unexpectedly
- Model behavior can vary across different versions
- Some reasoning tasks require specific prompting techniques
Resource Considerations
- Longer prompts consume more tokens and increase costs
- Complex few-shot examples impact context usage
- Multiple model calls for optimization increase expenses
- Testing requires sufficient sample sizes for statistical significance
Quality Requirements
- Validate prompts with domain-specific test cases
- Ensure prompts generalize beyond training examples
- Implement monitoring for prompt performance drift
- Regularly review and update prompts as requirements evolve
This skill provides the foundational patterns and methodologies for building production-ready prompt systems that consistently deliver high performance across diverse use cases and model types.
相关 Skills
网页构建器
by anthropics
面向复杂 claude.ai HTML artifact 开发,快速初始化 React + Tailwind CSS + shadcn/ui 项目并打包为单文件 HTML,适合需要状态管理、路由或多组件交互的页面。
✎ 在 claude.ai 里做复杂网页 Artifact 很省心,多组件、状态和路由都能顺手搭起来,React、Tailwind 与 shadcn/ui 组合效率高、成品也更精致。
前端设计
by anthropics
面向组件、页面、海报和 Web 应用开发,按鲜明视觉方向生成可直接落地的前端代码与高质感 UI,适合做 landing page、Dashboard 或美化现有界面,避开千篇一律的 AI 审美。
✎ 想把页面做得既能上线又有设计感,就用前端设计:组件到整站都能产出,难得的是能避开千篇一律的 AI 味。
网页应用测试
by anthropics
用 Playwright 为本地 Web 应用编写自动化测试,支持启动开发服务器、校验前端交互、排查 UI 异常、抓取截图与浏览器日志,适合调试动态页面和回归验证。
✎ 借助 Playwright 一站式验证本地 Web 应用前端功能,调 UI 时还能同步查看日志和截图,定位问题更快。
相关 MCP 服务
GitHub
编辑精选by GitHub
GitHub 是 MCP 官方参考服务器,让 Claude 直接读写你的代码仓库和 Issues。
✎ 这个参考服务器解决了开发者想让 AI 安全访问 GitHub 数据的问题,适合需要自动化代码审查或 Issue 管理的团队。但注意它只是参考实现,生产环境得自己加固安全。
Context7 文档查询
编辑精选by Context7
Context7 是实时拉取最新文档和代码示例的智能助手,让你告别过时资料。
✎ 它能解决开发者查找文档时信息滞后的问题,特别适合快速上手新库或跟进更新。不过,依赖外部源可能导致偶尔的数据延迟,建议结合官方文档使用。
by tldraw
tldraw 是让 AI 助手直接在无限画布上绘图和协作的 MCP 服务器。
✎ 这解决了 AI 只能输出文本、无法视觉化协作的痛点——想象让 Claude 帮你画流程图或白板讨论。最适合需要快速原型设计或头脑风暴的开发者。不过,目前它只是个基础连接器,你得自己搭建画布应用才能发挥全部潜力。