终端GPT
Shell Gpt
by bytesagain
Run AI-powered shell commands and prompts for faster terminal workflows. Use when generating scripts, comparing outputs, evaluating prompts.
安装
claude skill add --url github.com/openclaw/skills/tree/main/skills/bytesagain/shell-gpt文档
Shell Gpt
A command-line productivity tool powered by AI large language models like GPT-5, designed to help you accomplish terminal tasks faster and more efficiently. Log, track, and manage AI-related workflows including prompt evaluation, model benchmarking, cost tracking, and output analysis — all from a single CLI.
Inspired by TheR1D/shell_gpt (11,891+ GitHub stars).
Commands
-
configure— Set up or review configuration parameters for Shell Gpt. Run without arguments to view recent config entries, or pass key-value pairs to save new settings. Useful for persisting API keys, model preferences, or runtime flags. -
benchmark— Record and review benchmark results for model or prompt performance. Without arguments it shows recent benchmark logs; with arguments it saves a new benchmark entry with a timestamp. Great for tracking latency, throughput, or quality metrics over time. -
compare— Log side-by-side comparison notes between models, prompts, or outputs. View previous comparisons with no arguments, or save a new comparison record. Helps you keep a structured journal of A/B test results. -
prompt— Store and retrieve prompt templates or prompt-engineering notes. Run bare to see recent prompts, or pass text to save a new prompt entry. Essential for iterating on prompt design and keeping a prompt library. -
evaluate— Record evaluation scores or qualitative assessments of model outputs. Without arguments it displays recent evaluations; with arguments it logs a new evaluation. Use this to build a track record of output quality over time. -
fine-tune— Track fine-tuning jobs, hyperparameters, and results. Shows recent fine-tune logs when called alone, or saves a new entry when given input. Handy for documenting training runs alongside their outcomes. -
analyze— Log analysis observations about data, outputs, or model behavior. View past analyses or record a new one. Use this as a lightweight lab notebook for your AI experiments. -
cost— Track API costs, token usage expenses, and billing notes. Without arguments it shows recent cost entries; with input it saves a new cost record. Critical for keeping cloud AI spending under control. -
usage— Record and review API usage statistics such as token counts or request volumes. Displays recent usage entries when called bare, or logs new usage data when given arguments. Helps you stay within rate limits and quotas. -
optimize— Document optimization attempts, parameter tweaks, and their effects. Shows recent optimization logs or saves a new entry. Use this to track what you tried and what worked. -
test— Log test cases, test results, and validation outcomes. View recent test entries or record a new one. Perfect for building a regression test journal for your prompts or pipelines. -
report— Save and review summary reports or periodic write-ups. Without arguments it shows recent reports; with text it saves a new report entry. Ideal for weekly summaries or stakeholder updates. -
stats— Display summary statistics across all logged categories. Shows total entry counts per category, overall totals, and data directory disk usage. A quick dashboard for your Shell Gpt activity. -
export <fmt>— Export all logged data to a file in the specified format:json,csv, ortxt. The exported file is saved to the data directory and its size is reported. Use this for backups, data migration, or feeding logs into other tools. -
search <term>— Search across all log files for a given term (case-insensitive). Returns matching lines grouped by category. Helpful for finding a specific entry when you remember a keyword but not which command logged it. -
recent— Show the 20 most recent entries from the global activity history log. Gives you a quick at-a-glance view of what has been happening across all commands. -
status— Run a health check on the Shell Gpt installation. Reports the current version, data directory path, total log entries, disk usage, last activity timestamp, and overall status. Use this to verify the tool is working correctly. -
help— Display the full usage guide listing all available commands and their syntax. Also shows the current data directory path. -
version— Print the current Shell Gpt version string (shell-gpt v2.0.0).
When to Use
- Prompt engineering sessions — When you are iterating on prompt designs, use
promptto save drafts,evaluateto score outputs, andcompareto log A/B differences between prompt variants. - Model benchmarking and selection — When comparing multiple LLMs or model versions, use
benchmarkto record performance metrics,analyzeto note behavioral observations, andreportto write up your conclusions. - Cost and usage monitoring — When you need to keep cloud AI spending in check, use
costto log expenses,usageto track token consumption, andstatsto get an overview of activity volume. - Fine-tuning workflow tracking — When running fine-tuning jobs, use
fine-tuneto document hyperparameters and results,testto record validation outcomes, andoptimizeto log parameter adjustments. - Data export and auditing — When you need to share logs, create backups, or feed data into dashboards, use
exportto dump everything to JSON/CSV/TXT, andsearchto locate specific entries quickly.
Examples
# Save a new prompt template and review recent prompts
shell-gpt prompt "Summarize the following text in 3 bullet points:"
shell-gpt prompt
# Benchmark a model and check overall stats
shell-gpt benchmark "gpt-4o: 320ms avg latency, 92% accuracy on eval set"
shell-gpt stats
# Log API costs and export all data as CSV
shell-gpt cost "2026-03-18: $4.27 for 1.2M tokens on gpt-4o"
shell-gpt export csv
# Compare two models side by side
shell-gpt compare "gpt-4o vs claude-opus: gpt-4o faster, claude-opus more detailed"
# Search across all logs for a keyword and check system health
shell-gpt search "latency"
shell-gpt status
Output
Returns reports and log entries to stdout. All data is persisted to per-command log files in the data directory. Redirect output to a file with shell-gpt <command> > output.txt.
Configuration
Set the SHELL_GPT_DIR environment variable to change the data directory. Default: ~/.local/share/shell-gpt/. All log files and exports are stored in this directory.
Powered by BytesAgain | bytesagain.com Feedback & Feature Requests: https://bytesagain.com/feedback
相关 Skills
技能工坊
by anthropics
覆盖 Skill 从创建到迭代优化全流程:起草能力、补测试提示、跑评测与基准方差分析,并持续改写内容和描述,提升效果与触发准确率。
✎ 技能工坊把技能从创建、迭代到评测串成闭环,方差分析加描述优化,特别适合把触发准确率打磨得更稳。
表格处理
by anthropics
围绕 .xlsx、.xlsm、.csv、.tsv 做读写、修复、清洗、格式整理、公式计算与格式转换,适合修改现有表格、生成新报表或把杂乱数据整理成交付级电子表格。
✎ 做 Excel/CSV 相关任务很省心,能直接读写、修复、清洗和格式转换,尤其擅长把乱七八糟的表格整理成交付级文件。
Word文档
by anthropics
覆盖Word/.docx文档的创建、读取、编辑与重排,适合生成报告、备忘录、信函和模板,也能处理目录、页眉页脚、页码、图片替换、查找替换、修订批注及内容提取整理。
✎ 搞定 .docx 的创建、改写与精排版,目录、批量替换、批注修订和图片更新都能自动化,做正式文档尤其省心。
相关 MCP 服务
文件系统
编辑精选by Anthropic
Filesystem 是 MCP 官方参考服务器,让 LLM 安全读写本地文件系统。
✎ 这个服务器解决了让 Claude 直接操作本地文件的痛点,比如自动整理文档或生成代码文件。适合需要自动化文件处理的开发者,但注意它只是参考实现,生产环境需自行加固安全。
by wonderwhy-er
Desktop Commander 是让 AI 直接执行终端命令、管理文件和进程的 MCP 服务器。
✎ 这工具解决了 AI 无法直接操作本地环境的痛点,适合需要自动化脚本调试或文件批量处理的开发者。它能让你用自然语言指挥终端,但权限控制需谨慎,毕竟让 AI 执行 rm -rf 可不是闹着玩的。
EdgarTools
编辑精选by dgunning
EdgarTools 是无需 API 密钥即可解析 SEC EDGAR 财报的开源 Python 库。
✎ 这个工具解决了金融数据获取的痛点——直接让 AI 读取结构化财报,比如让 Claude 分析苹果的 10-K 文件。适合量化分析师或金融开发者快速构建数据管道。但注意,它依赖 SEC 网站稳定性,高峰期可能延迟。