expertpack
by brianhearn
Work with ExpertPacks — structured knowledge packs for AI agents. Use when: (1) Loading/consuming an ExpertPack as agent context, (2) Creating or hydrating a new ExpertPack from scratch, (3) Chunking a pack for RAG deployment, (4) Backing up/exporting an OpenClaw agent's workspace into an ExpertPack. Triggers on: 'expertpack', 'expert pack', 'esoteric knowledge', 'knowledge pack', 'pack hydration', 'backup to expertpack', 'export agent knowledge'. For EK ratio measurement and quality evals, install the separate expertpack-eval skill.
安装
claude skill add --url github.com/openclaw/skills/tree/main/skills/brianhearn/expertpack必需命令行工具
python3文档
ExpertPack
Structured knowledge packs for AI agents. Maximize the knowledge your AI is missing.
Learn more: expertpack.ai · GitHub · Schema docs
Full schemas: /path/to/ExpertPack/schemas/ in the repo (core.md, person.md, product.md, process.md, composite.md, eval.md)
Pack Location
Default directory: ~/expertpacks/. Check there first, fall back to current workspace. Users can override by specifying a path.
Actions
1. Load / Consume a Pack
- Read
manifest.yaml— identify type, version, context tiers - Read
overview.md— understand what the pack covers - Load all Tier 1 (always) files into session context
- For queries: search Tier 2 (searchable) files via RAG or
_index.mdnavigation - Load Tier 3 (on-demand) only on explicit request (verbatim transcripts, training data)
OpenClaw RAG config — add to openclaw.json:
{
"agents": {
"defaults": {
"memorySearch": {
"extraPaths": ["path/to/pack/.chunks"],
"chunking": { "tokens": 500, "overlap": 0 },
"query": {
"hybrid": {
"enabled": true,
"mmr": { "enabled": true, "lambda": 0.7 },
"temporalDecay": { "enabled": false }
}
}
}
}
}
}
For detailed platform integration (Cursor, Claude Code, custom APIs, direct context window): read {skill_dir}/references/consumption.md.
2. Create / Hydrate a Pack
- Determine pack type: person, product, process, or composite
- Read
{skill_dir}/references/schemas.mdfor structural requirements - Scaffold the directory structure per the type schema
- Create
manifest.yamlandoverview.md(both required) - Populate content using EK-aware hydration:
- Blind-probe each extracted fact before filing
- Full treatment for EK content (the model can't produce it)
- Compressed scaffolding for GK content (the model already knows it)
- Skip content with zero EK value
- Add retrieval layers:
_index.mdper directory,summaries/,propositions/,glossary.md - Add
sources/_coverage.mddocumenting what was researched
For full hydration methodology, EK triage process, and source prioritization: read {skill_dir}/references/hydration.md.
3. Chunk for RAG
Run the schema-aware chunker:
python3 {skill_dir}/scripts/chunk.py --pack <pack-path> --output <pack-path>/.chunks
- Respects
##headers, lead summaries, proposition groups,<!-- refresh -->metadata - Each output file = one semantically coherent chunk
- Point OpenClaw RAG at
.chunks/with overlap=0
Why this matters: Schema-aware chunking produced +9.4% correctness and -52% tokens vs. generic chunking in controlled experiments. It's the single highest-impact consumption optimization.
4. Measure EK Ratio & Run Quality Evals
For EK ratio measurement (blind probing) and automated quality evals, install the companion skill:
clawhub install expertpack-eval
See expertpack-eval for full details on EK measurement, eval runner, and the improvement loop.
5. Backup / Export OpenClaw → ExpertPack
Export an OpenClaw agent's accumulated knowledge into a structured ExpertPack composite.
Step 1 — Scan:
python3 {skill_dir}/scripts/scan.py --workspace <workspace-path> --output /tmp/ep-scan.json
Review the scan output with the user. It proposes pack assignments (agent, person, product, process) with confidence scores. Flag ambiguous classifications for user decision.
Step 2 — Distill (repeat per pack):
python3 {skill_dir}/scripts/distill.py \
--scan /tmp/ep-scan.json \
--pack <type:slug> \
--output <export-dir>/packs/<slug>/
- Distill, don't copy — target 10-20% volume of raw state
- Strips secrets automatically (API keys, tokens, passwords)
- Deduplicates, prefers newest for conflicts
Step 3 — Compose:
python3 {skill_dir}/scripts/compose.py \
--scan /tmp/ep-scan.json \
--export-dir <export-dir>/
Generates composite manifest and overview.
Step 4 — Validate:
python3 {skill_dir}/scripts/validate.py --export-dir <export-dir>/
Checks: required files exist, manifest fields valid, no secrets leaked, file sizes within guidelines, cross-references resolve.
Step 5 — Review & ship. Present validation report to user. They decide whether to commit/push.
Critical rules:
- Never include secrets in the export
- Never modify the live workspace — all output goes to the export directory
- Flag personal information for access tier review
- Default user-specific content to
privateaccess
相关 Skills
Claude接口
by anthropics
面向接入 Claude API、Anthropic SDK 或 Agent SDK 的开发场景,自动识别项目语言并给出对应示例与默认配置,快速搭建 LLM 应用。
✎ 想把Claude能力接进应用或智能体,用claude-api上手快、兼容Anthropic与Agent SDK,集成路径清晰又省心
提示工程专家
by alirezarezvani
覆盖Prompt优化、Few-shot设计、结构化输出、RAG评测与Agent工作流编排,适合分析token成本、评估LLM输出质量,并搭建可落地的AI智能体系统。
✎ 把提示优化、LLM评测到RAG与智能体设计串成一套方法,适合想系统提升AI开发效率的人。
智能体流程设计
by alirezarezvani
面向生产级多 Agent 编排,梳理顺序、并行、分层、事件驱动、共识五种工作流设计,覆盖 handoff、状态管理、容错重试、上下文预算与成本优化,适合搭建复杂 AI 协作系统。
✎ 帮你把多智能体流程设计、编排和自动化统一起来,复杂工作流也能更稳地落地,适合追求强控制力的团队。
相关 MCP 服务
顺序思维
编辑精选by Anthropic
Sequential Thinking 是让 AI 通过动态思维链解决复杂问题的参考服务器。
✎ 这个服务器展示了如何让 Claude 像人类一样逐步推理,适合开发者学习 MCP 的思维链实现。但注意它只是个参考示例,别指望直接用在生产环境里。
知识图谱记忆
编辑精选by Anthropic
Memory 是一个基于本地知识图谱的持久化记忆系统,让 AI 记住长期上下文。
✎ 帮 AI 和智能体补上“记不住”的短板,用本地知识图谱沉淀长期上下文,连续对话更聪明,数据也更可控。
PraisonAI
编辑精选by mervinpraison
PraisonAI 是一个支持自反思和多 LLM 的低代码 AI 智能体框架。
✎ 如果你需要快速搭建一个能 24/7 运行的 AI 智能体团队来处理复杂任务(比如自动研究或代码生成),PraisonAI 的低代码设计和多平台集成(如 Telegram)让它上手极快。但作为非官方项目,它的生态成熟度可能不如 LangChain 等主流框架,适合愿意尝鲜的开发者。