微信读书推荐器

weread-reading-recommender

by cengsin

Use this skill when the user wants to export local WeRead records, normalize WeRead data, analyze reading preferences from WeRead history, or get book recommendations grounded in WeRead reading behavior and a current learning goal.

3.9k数据与存储未扫描2026年3月23日

安装

claude skill add --url github.com/openclaw/skills/tree/main/skills/cengsin/weread-reading-recommender

文档

WeRead Reading Recommender

Overview

This is a local-first skill for exporting 微信读书 (WeRead) records from a cookie stored on the user's machine, normalizing those records into a recommendation-friendly JSON file, and using that data to analyze reading preferences or recommend what to read next.

Use this skill when the user wants to:

  • 根据微信读书记录推荐书
  • 分析自己的阅读偏好或阅读画像
  • 结合“最近想学的主题”与微信读书历史一起做推荐
  • 导出、刷新、归一化本地微信读书数据

Trigger Cases

Activate this skill for requests like:

  • “根据我的微信读书记录推荐书”
  • “分析我的阅读偏好”
  • “我最近想系统学 AI Agent,结合微信读书记录推荐 5 本书”
  • “帮我导出 / 刷新 / 归一化微信读书数据”
  • “基于我的阅读历史,推荐下一本最适合现在读的书”
  • “分析我的阅读偏好,并给我 3 本稳妥推荐 + 2 本探索推荐”

Workflow

Follow this sequence:

  1. Check whether a normalized JSON file already exists.
  2. If normalized data is missing, or the user explicitly wants fresh data, check whether a local WeRead cookie is already available.
  3. Look for a local cookie source in this order:
    • a cookie file path explicitly provided by the user
    • WEREAD_COOKIE
    • another env var name passed through --env-var
  4. If no local cookie source exists, ask the user to set one locally and stop there. Do not tell the user to edit SKILL.md.
  5. If a local cookie source exists, run the export script.
  6. Run the normalize script on the raw export.
  7. Read the normalized JSON and identify strong signals:
    • high-engagement books
    • recent books
    • unfinished books with momentum
    • repeated categories or lists
  8. If the user provides a current goal, weight goal fit first.
  9. If the user does not provide a goal, produce a reading-profile summary plus safe and exploratory recommendations.

Recommendation Guidance

When the user provides a current goal, weight approximately:

  • 60% goal fit
  • 40% history fit

When the user provides no goal, weight approximately:

  • 70% history fit
  • 20% recency
  • 10% exploration/diversity

For each recommendation, explain:

  • why it fits the user's current goal or history
  • which past books it resembles
  • what gap it fills
  • whether it is a safe pick or an exploration pick
  • whether it is a good fit right now

Suggested response structure:

  • 阅读画像 / Reading profile
  • 推荐结果 / Recommendations
  • 为什么适合现在 / Why now
  • 暂缓推荐 / Skip for now (optional)

Local Data Workflow

1. Check local cookie availability first

Before asking the user to set anything, first check whether a local cookie is already available through:

  • a cookie file path the user provided
  • WEREAD_COOKIE
  • another env var name passed through --env-var

If none of these exist, ask the user to set the cookie locally, then continue.

2. Export raw WeRead data

If a local cookie is already available, export directly:

bash
python3 scripts/export_weread.py --out data/weread-raw.json

Optional variants:

bash
python3 scripts/export_weread.py --cookie-file ~/.config/weread.cookie --out data/weread-raw.json
python3 scripts/export_weread.py --env-var WEREAD_COOKIE --include-book-info --detail-limit 50 --out data/weread-raw.json

If the user does need to set one manually, keep it local. For example:

bash
export WEREAD_COOKIE='wr_skey=...; wr_vid=...; ...'

3. Normalize the raw export

bash
python3 scripts/normalize_weread.py --input data/weread-raw.json --output data/weread-normalized.json

4. Use the normalized file for recommendation turns

After normalization, this skill should reason primarily from the normalized JSON, not from a live cookie session, unless the user explicitly asks for a refresh.

Security Boundary

This skill is local-first. Enforce these rules:

  • Cookie is for local use only.
  • Never write the cookie into SKILL.md, scripts, assets, logs, or exported JSON.
  • Never echo the cookie back in responses.
  • Prefer checking existing local cookie sources before asking the user to set one again.
  • Do not rely on CookieCloud or any third-party cookie sync service by default.
  • Do not suggest remote cookie hosting as the normal path.
  • Recommendation work should use the normalized JSON whenever possible.

Files

Use these project files as the main references:

  • scripts/export_weread.py
  • scripts/normalize_weread.py
  • references/data-schema.md
  • references/privacy-model.md
  • references/recommendation-rubric.md
  • assets/sample-weread-raw.json
  • assets/sample-weread-normalized.json

Example Requests

  • 结合我的微信读书记录,我最近想系统学 AI Agent,推荐 5 本书
  • 基于我的阅读历史,推荐下一本最适合现在读的书
  • 分析我的阅读偏好,并给我 3 本稳妥推荐 + 2 本探索推荐
  • 帮我刷新微信读书数据,然后按最近在读主题推荐下一批书

相关 Skills

数据库建模

by alirezarezvani

Universal
热门

把需求梳理成关系型数据库表结构,自动生成迁移脚本、TypeScript/Python 类型、种子数据、RLS 策略和索引方案,适合多租户、审计追踪、软删除等后端建模与 Schema 评审场景。

把数据库结构设计、ER图梳理和SQL建模放到一处,复杂业务也能快速统一数据模式,少走不少返工弯路。

数据与存储
未扫描9.8k

资深数据科学家

by alirezarezvani

Universal
热门

覆盖实验设计、特征工程、预测建模、因果推断与模型评估,适合用 Python/R/SQL 做 A/B 测试、时序分析和生产级 ML 落地,支撑数据驱动决策。

从 A/B 测试、因果分析到预测建模一条龙搞定,既有硬核统计方法也懂业务沟通,特别适合把数据结论真正落地。

数据与存储
未扫描9.8k

数据库设计

by alirezarezvani

Universal
热门

聚焦数据库 Schema 设计与演进,自动检查规范化、数据类型、约束和索引问题,生成 ERD,并为零停机迁移、数据变更和回滚提供可执行方案。

专注数据库设计与数据建模,帮你快速理清表结构和关系,减少后期返工,SQL 落地也更顺手。

数据与存储
未扫描9.8k

相关 MCP 服务

by Anthropic

热门

PostgreSQL 是让 Claude 直接查询和管理你的数据库的 MCP 服务器。

这个服务器解决了开发者需要手动编写 SQL 查询的痛点,特别适合数据分析师或后端开发者快速探索数据库结构。不过,由于是参考实现,生产环境使用前务必评估安全风险,别指望它能处理复杂事务。

数据与存储
83.1k

SQLite 数据库

编辑精选

by Anthropic

热门

SQLite 是让 AI 直接查询本地数据库进行数据分析的 MCP 服务器。

这个服务器解决了 AI 无法直接访问 SQLite 数据库的问题,适合需要快速分析本地数据集的开发者。不过,作为参考实现,它可能缺乏生产级的安全特性,建议在受控环境中使用。

数据与存储
83.1k

by Firecrawl

热门

Firecrawl 是让 AI 直接抓取网页并提取结构化数据的 MCP 服务器。

它解决了手动写爬虫的麻烦,让 Claude 能直接访问动态网页内容。最适合需要实时数据的研究者或开发者,比如监控竞品价格或抓取新闻。但要注意,它依赖第三方 API,可能涉及隐私和成本问题。

数据与存储
6.0k

评论