终端看板

Sampler

by bytesagain3

Tool for shell commands execution, visualization and alerting. Configured with a simple YAML file. terminal-dashboard, go, alerting, charts, cmd.

3.7k效率与工作流未扫描2026年3月23日

安装

claude skill add --url github.com/openclaw/skills/tree/main/skills/bytesagain3/terminal-dashboard

文档

Terminal Dashboard

Terminal Dashboard v2.0.0 — a data toolkit for building data pipelines and tracking data operations from the command line. Ingest, transform, query, filter, aggregate, and visualize your data — all logged locally with timestamps for full traceability.

Why Terminal Dashboard?

  • Works entirely offline — your data never leaves your machine
  • Simple command-line interface, no GUI needed
  • Timestamped logging for every operation
  • Export to JSON, CSV, or plain text anytime
  • Automatic history and activity tracking
  • Searchable records across all data pipeline stages

Getting Started

bash
# See all available commands
terminal-dashboard help

# Check current health status
terminal-dashboard status

# View summary statistics
terminal-dashboard stats

Commands

Data Pipeline Commands

Each command works in two modes: run without arguments to view recent entries, or pass input to record a new entry.

CommandDescription
terminal-dashboard ingest <input>Record data ingestion events (file imports, API pulls, stream captures)
terminal-dashboard transform <input>Log data transformations (format conversions, cleaning steps, enrichments)
terminal-dashboard query <input>Record queries executed (SQL, API calls, search operations)
terminal-dashboard filter <input>Log filter operations (row filtering, column selection, deduplication)
terminal-dashboard aggregate <input>Record aggregation operations (group-by, rollups, summaries)
terminal-dashboard visualize <input>Log visualization outputs (charts generated, dashboards updated)
terminal-dashboard export <input>Record export operations (file writes, API pushes, report generation)
terminal-dashboard sample <input>Log sampling operations (random samples, stratified picks, head/tail)
terminal-dashboard schema <input>Record schema operations (schema detection, validation rules, migrations)
terminal-dashboard validate <input>Log validation results (data quality checks, constraint tests, anomalies)
terminal-dashboard pipeline <input>Record pipeline operations (end-to-end runs, DAG executions, orchestration)
terminal-dashboard profile <input>Log profiling results (data profiling, column stats, distribution analysis)

Utility Commands

CommandDescription
terminal-dashboard statsShow summary statistics across all log categories
terminal-dashboard export <fmt>Export all data (formats: json, csv, txt)
terminal-dashboard search <term>Search across all entries for a keyword
terminal-dashboard recentShow the 20 most recent history entries
terminal-dashboard statusHealth check — version, data dir, entry count, disk usage
terminal-dashboard helpShow the built-in help message
terminal-dashboard versionPrint version (v2.0.0)

Data Storage

All data is stored locally in ~/.local/share/terminal-dashboard/. Structure:

  • ingest.log, transform.log, query.log, etc. — one log file per command, pipe-delimited (timestamp|value)
  • history.log — unified activity log across all commands
  • export.json / export.csv / export.txt — generated export files

Each entry is stored as YYYY-MM-DD HH:MM|<input>. Use export to back up your data anytime.

Requirements

  • Bash 4+ (uses set -euo pipefail)
  • Standard Unix utilities (date, wc, du, tail, grep, sed, cat)
  • No external dependencies or internet access needed

When to Use

  1. Data pipeline logging — Track every step of your ETL/ELT pipeline from ingestion through transformation to export, creating a complete audit trail
  2. Data quality monitoring — Use validate and profile to record data quality checks and catch anomalies before they reach production
  3. Schema change tracking — Log schema migrations and validation rules so you always know what changed and when
  4. Ad-hoc analysis journaling — Record queries, filters, and aggregations during exploratory analysis so you can reproduce your findings later
  5. Pipeline debugging — When a data pipeline breaks, search through ingest, transform, and export logs to pinpoint where things went wrong

Examples

bash
# Record a data ingestion event
terminal-dashboard ingest "Loaded 2.4M rows from sales_2024.csv into staging"

# Log a transformation step
terminal-dashboard transform "Normalized phone numbers, deduplicated by email — 12k dupes removed"

# Record a query
terminal-dashboard query "SELECT region, SUM(revenue) FROM sales GROUP BY region — 8 rows returned"

# Log a validation check
terminal-dashboard validate "Schema check passed: all 47 columns match expected types"

# Record a pipeline run
terminal-dashboard pipeline "Daily ETL completed: ingest→clean→aggregate→export in 4m 23s"

# Export everything to JSON
terminal-dashboard export json

# Search logs for a dataset
terminal-dashboard search "sales_2024"

Output

All commands output to stdout. Redirect to a file if needed:

bash
terminal-dashboard stats > pipeline-report.txt
terminal-dashboard export csv

Configuration

Set TERMINAL_DASHBOARD_DIR environment variable to override the default data directory (~/.local/share/terminal-dashboard/).


Powered by BytesAgain | bytesagain.com | hello@bytesagain.com

相关 Skills

表格处理

by anthropics

Universal
热门

围绕 .xlsx、.xlsm、.csv、.tsv 做读写、修复、清洗、格式整理、公式计算与格式转换,适合修改现有表格、生成新报表或把杂乱数据整理成交付级电子表格。

做 Excel/CSV 相关任务很省心,能直接读写、修复、清洗和格式转换,尤其擅长把乱七八糟的表格整理成交付级文件。

效率与工作流
未扫描109.6k

PDF处理

by anthropics

Universal
热门

遇到 PDF 读写、文本表格提取、合并拆分、旋转加水印、表单填写或加解密时直接用它,也能提取图片、生成新 PDF,并把扫描件通过 OCR 变成可搜索文档。

PDF杂活别再来回切工具了,文本表格提取、合并拆分到OCR识别一次搞定,连扫描件也能变可搜索。

效率与工作流
未扫描109.6k

Word文档

by anthropics

Universal
热门

覆盖Word/.docx文档的创建、读取、编辑与重排,适合生成报告、备忘录、信函和模板,也能处理目录、页眉页脚、页码、图片替换、查找替换、修订批注及内容提取整理。

搞定 .docx 的创建、改写与精排版,目录、批量替换、批注修订和图片更新都能自动化,做正式文档尤其省心。

效率与工作流
未扫描109.6k

相关 MCP 服务

文件系统

编辑精选

by Anthropic

热门

Filesystem 是 MCP 官方参考服务器,让 LLM 安全读写本地文件系统。

这个服务器解决了让 Claude 直接操作本地文件的痛点,比如自动整理文档或生成代码文件。适合需要自动化文件处理的开发者,但注意它只是参考实现,生产环境需自行加固安全。

效率与工作流
82.9k

by wonderwhy-er

热门

Desktop Commander 是让 AI 直接执行终端命令、管理文件和进程的 MCP 服务器。

这工具解决了 AI 无法直接操作本地环境的痛点,适合需要自动化脚本调试或文件批量处理的开发者。它能让你用自然语言指挥终端,但权限控制需谨慎,毕竟让 AI 执行 rm -rf 可不是闹着玩的。

效率与工作流
5.8k

EdgarTools

编辑精选

by dgunning

热门

EdgarTools 是无需 API 密钥即可解析 SEC EDGAR 财报的开源 Python 库。

这个工具解决了金融数据获取的痛点——直接让 AI 读取结构化财报,比如让 Claude 分析苹果的 10-K 文件。适合量化分析师或金融开发者快速构建数据管道。但注意,它依赖 SEC 网站稳定性,高峰期可能延迟。

效率与工作流
1.9k

评论