ondeckllm

by canonflip-git

>

View Chinese version with editor review

安装

npm install -g ondeckllm

文档

OnDeckLLM — AI Model Lineup Manager

Prerequisites

bash
npm install -g ondeckllm

Verify: ondeckllm --help or check install with npm list -g ondeckllm.

What It Does

OnDeckLLM is a localhost web dashboard that:

  • Auto-discovers LLM providers from OpenClaw config (~/.openclaw/openclaw.json)
  • Manages a batting-order priority list for model routing (primary + fallbacks)
  • Tests provider health and latency
  • Syncs model lineup back to OpenClaw config with one click
  • Tracks session costs (JSONL usage log + Chart.js)
  • Supports Anthropic, OpenAI, Google AI, Groq, xAI/Grok, Ollama (local + remote), Mistral, DeepSeek, Together, OpenRouter

Starting the Dashboard

bash
# Default port 3900
ondeckllm

# Custom port
PORT=3901 ondeckllm

The dashboard runs at http://localhost:3900 (or custom port).

As a Background Service

Use the helper script to check status or start OnDeckLLM:

bash
node scripts/status.js

Output: JSON with running (bool), port, url, and pid if active.

Agent Workflow

Check if OnDeckLLM is running

bash
node scripts/status.js

Open the dashboard for the user

Direct them to http://localhost:3900 (or the configured port/URL).

Provider management

OnDeckLLM reads provider config from ~/.openclaw/openclaw.json automatically. Changes made in the dashboard sync back to OpenClaw config. No separate API or CLI commands needed — it's a web UI tool.

Configuration

OnDeckLLM stores its data in ~/.ondeckllm/:

  • config.json — provider settings, port, Ollama URL
  • usage.jsonl — cost tracking log
  • profiles/ — saved batting-order profiles

Remote Ollama

To connect to a remote Ollama instance, configure in the dashboard UI: Settings → Ollama → Remote URL (e.g., http://192.168.55.80:11434)

Links