Datasets
by bytesagain
🤗 The largest hub of ready-to-use dataset-loader for AI models with fast, easy-to-use and efficient data m dataset-loader, python, ai.
安装
claude skill add --url github.com/openclaw/skills/tree/main/skills/bytesagain/dataset-loader文档
Dataset Loader
A content creation and management toolkit for drafting, editing, optimizing, and scheduling content from the command line. All operations are logged with timestamps and stored locally.
Commands
Content Operations
Each content command works in two modes: run without arguments to view recent entries, or pass input to record a new entry.
| Command | Description |
|---|---|
dataset-loader draft <input> | Draft content — record a new draft or view recent ones |
dataset-loader edit <input> | Edit content — record an edit or view recent ones |
dataset-loader optimize <input> | Optimize content — record an optimization or view recent ones |
dataset-loader schedule <input> | Schedule content — record a schedule entry or view recent ones |
dataset-loader hashtags <input> | Generate hashtags — record hashtags or view recent ones |
dataset-loader hooks <input> | Create hooks — record a hook or view recent ones |
dataset-loader cta <input> | Create call-to-action — record a CTA or view recent ones |
dataset-loader rewrite <input> | Rewrite content — record a rewrite or view recent ones |
dataset-loader translate <input> | Translate content — record a translation or view recent ones |
dataset-loader tone <input> | Adjust tone — record a tone adjustment or view recent ones |
dataset-loader headline <input> | Create headline — record a headline or view recent ones |
dataset-loader outline <input> | Create outline — record an outline or view recent ones |
Utility Commands
| Command | Description |
|---|---|
dataset-loader stats | Show summary statistics — entry counts per category, total entries, disk usage |
dataset-loader export <fmt> | Export all data to a file (formats: json, csv, txt) |
dataset-loader search <term> | Search all log files for a term (case-insensitive) |
dataset-loader recent | Show last 20 entries from activity history |
dataset-loader status | Health check — version, data directory, entry count, disk usage, last activity |
dataset-loader help | Show available commands |
dataset-loader version | Show version (v2.0.0) |
Data Storage
All data is stored locally at ~/.local/share/dataset-loader/:
- Each content command writes to its own log file (e.g.,
draft.log,edit.log,hashtags.log) - Entries are stored as
timestamp|valuepairs (pipe-delimited) - All actions are tracked in
history.logwith timestamps - Export generates files in the data directory (
export.json,export.csv, orexport.txt)
Requirements
- Bash (with
set -euo pipefail) - Standard Unix utilities:
date,wc,du,grep,tail,cat,sed - No external dependencies or API keys required
When to Use
- To log and track content creation workflows (draft, edit, optimize, schedule)
- To maintain a searchable history of content operations
- To manage hashtags, hooks, CTAs, and headlines in a structured way
- To export accumulated content records in JSON, CSV, or plain text format
- As part of larger content automation pipelines
Examples
# Draft new content
dataset-loader draft "Blog post about AI trends in 2026"
# View recent edits
dataset-loader edit
# Record hashtags
dataset-loader hashtags "#AI #MachineLearning #DataScience"
# Create a headline
dataset-loader headline "5 Ways AI Is Changing Data Science"
# Search across all logs
dataset-loader search "AI"
# Export everything as CSV
dataset-loader export csv
# Check statistics
dataset-loader stats
# View recent activity
dataset-loader recent
# Health check
dataset-loader status
Powered by BytesAgain | bytesagain.com | hello@bytesagain.com 💬 Feedback & Feature Requests: https://bytesagain.com/feedback
相关 Skills
迁移架构师
by alirezarezvani
为数据库、API 与基础设施迁移制定分阶段零停机方案,提前校验兼容性与风险,生成回滚策略、验证关卡和时间线,适合复杂系统平滑切换。
✎ 做数据库与存储迁移时,用它统一梳理表结构和数据搬迁流程,架构视角更完整,复杂迁移也更稳。
数据库建模
by alirezarezvani
把需求梳理成关系型数据库表结构,自动生成迁移脚本、TypeScript/Python 类型、种子数据、RLS 策略和索引方案,适合多租户、审计追踪、软删除等后端建模与 Schema 评审场景。
✎ 把数据库结构设计、ER图梳理和SQL建模放到一处,复杂业务也能快速统一数据模式,少走不少返工弯路。
资深数据工程师
by alirezarezvani
聚焦生产级数据工程,覆盖 ETL/ELT、批处理与流式管道、数据建模、Airflow/dbt/Spark 优化和数据质量治理,适合设计数据架构、搭建现代数据栈与排查性能问题。
✎ 复杂数据管道、ETL/ELT 和治理难题交给它,凭 Spark、Airflow、dbt 等现代数据栈经验,能更稳地搭起可扩展的数据基础设施。
相关 MCP 服务
PostgreSQL 数据库
编辑精选by Anthropic
PostgreSQL 是让 Claude 直接查询和管理你的数据库的 MCP 服务器。
✎ 这个服务器解决了开发者需要手动编写 SQL 查询的痛点,特别适合数据分析师或后端开发者快速探索数据库结构。不过,由于是参考实现,生产环境使用前务必评估安全风险,别指望它能处理复杂事务。
SQLite 数据库
编辑精选by Anthropic
SQLite 是让 AI 直接查询本地数据库进行数据分析的 MCP 服务器。
✎ 这个服务器解决了 AI 无法直接访问 SQLite 数据库的问题,适合需要快速分析本地数据集的开发者。不过,作为参考实现,它可能缺乏生产级的安全特性,建议在受控环境中使用。
Firecrawl 智能爬虫
编辑精选by Firecrawl
Firecrawl 是让 AI 直接抓取网页并提取结构化数据的 MCP 服务器。
✎ 它解决了手动写爬虫的麻烦,让 Claude 能直接访问动态网页内容。最适合需要实时数据的研究者或开发者,比如监控竞品价格或抓取新闻。但要注意,它依赖第三方 API,可能涉及隐私和成本问题。