io.github.Anarkitty1/semantic-frame

编码与调试

by anarkitty1

面向数值数据的高效语义压缩,显著节省 token,最高可实现 95% 以上 token reduction。

什么是 io.github.Anarkitty1/semantic-frame

面向数值数据的高效语义压缩,显著节省 token,最高可实现 95% 以上 token reduction。

README

Semantic Frame

<!-- mcp-name: io.github.Anarkitty1/semantic-frame -->

MCP Registry PyPI version License: MIT

Token-efficient semantic compression for numerical data.

Semantic Frame converts raw numerical data (NumPy, Pandas, Polars) into natural language descriptions optimized for LLM consumption. Instead of sending thousands of data points to an AI agent, send a 50-word semantic summary.

The Problem

LLMs are terrible at arithmetic. When you send raw data like [100, 102, 99, 101, 500, 100, 98] to GPT-4 or Claude:

  • Token waste: 1000 data points = ~2000 tokens
  • Hallucination risk: LLMs guess trends instead of calculating them
  • Context overflow: Large datasets fill the context window

The Solution

Semantic Frame provides deterministic analysis using NumPy, then translates results into token-efficient narratives:

python
from semantic_frame import describe_series
import pandas as pd

data = pd.Series([100, 102, 99, 101, 500, 100, 98])
print(describe_series(data, context="Server Latency (ms)"))

Output:

code
The Server Latency (ms) data shows a flat/stationary pattern with stable
variability. 1 anomaly detected at index 4 (value: 500.00).
Baseline: 100.00 (range: 98.00-500.00).

Result: 95%+ token reduction, zero hallucination risk.

Installation

bash
pip install semantic-frame

Or with uv:

bash
uv add semantic-frame

🤖 Claude Integration (MCP)

Semantic Frame is available on the official MCP Registry, enabling direct integration with Claude.

Claude Code CLI:

bash
claude mcp add semantic-frame

Claude Desktop - Add to your claude_desktop_config.json:

macOS: ~/Library/Application Support/Claude/claude_desktop_config.json Windows: %APPDATA%\Claude\claude_desktop_config.json

json
{
  "mcpServers": {
    "semantic-frame": {
      "command": "uvx",
      "args": ["--from", "semantic-frame[mcp]", "semantic-frame-mcp"]
    }
  }
}

Once configured, Claude can use these tools:

  • describe_data - Analyze a single data series
  • describe_batch - Analyze multiple series at once
  • describe_json - Get structured JSON output

Quick Start

Analyze a Series

python
from semantic_frame import describe_series
import numpy as np

# Works with NumPy arrays
data = np.array([10, 20, 30, 40, 50, 60, 70, 80, 90, 100])
result = describe_series(data, context="Daily Sales")
print(result)
# "The Daily Sales data shows a rapidly rising pattern with moderate variability..."

Analyze a DataFrame

python
from semantic_frame import describe_dataframe
import pandas as pd

df = pd.DataFrame({
    'cpu': [40, 42, 41, 95, 40, 41],
    'memory': [60, 61, 60, 60, 61, 60],
})

results = describe_dataframe(df, context="Server Metrics")
print(results['cpu'].narrative)
# "The Server Metrics - cpu data shows a flat/stationary pattern..."

Get Structured Output

python
result = describe_series(data, output="full")

print(result.trend)          # TrendState.RISING_SHARP
print(result.volatility)     # VolatilityState.MODERATE
print(result.anomalies)      # [AnomalyInfo(index=4, value=500.0, z_score=4.2)]
print(result.compression_ratio)  # 0.95

JSON Output for APIs

python
result = describe_series(data, output="json")
# Returns dict ready for JSON serialization

Supported Data Types

  • NumPy: np.ndarray
  • Pandas: pd.Series, pd.DataFrame
  • Polars: pl.Series, pl.DataFrame
  • Python: list

Analysis Features

FeatureMethodOutput
TrendLinear regression slopeRISING_SHARP, RISING_STEADY, FLAT, FALLING_STEADY, FALLING_SHARP
VolatilityCoefficient of variationCOMPRESSED, STABLE, MODERATE, EXPANDING, EXTREME
AnomaliesZ-score / IQR adaptiveIndex, value, z-score for each outlier
SeasonalityAutocorrelationNONE, WEAK, MODERATE, STRONG
DistributionSkewness + KurtosisNORMAL, LEFT_SKEWED, RIGHT_SKEWED, BIMODAL, UNIFORM
Step ChangeBaseline shift detectionNONE, STEP_UP, STEP_DOWN
Data QualityMissing value %PRISTINE, GOOD, SPARSE, FRAGMENTED

LLM Integration

System Prompt Injection

python
from semantic_frame.interfaces import format_for_system_prompt

result = describe_series(data, output="full")
prompt = format_for_system_prompt(result)
# Returns formatted context block for system prompts

LangChain Tool Output

python
from semantic_frame.interfaces import format_for_langchain

output = format_for_langchain(result)
# {"output": "narrative...", "metadata": {...}}

Multi-Column Agent Context

python
from semantic_frame.interfaces import create_agent_context

results = describe_dataframe(df)
context = create_agent_context(results)
# Combined narrative for all columns with attention flags

Framework Integrations

Anthropic Claude (Native Tool Use)

bash
pip install semantic-frame[anthropic]
python
import anthropic
from semantic_frame.integrations.anthropic import get_anthropic_tool, handle_tool_call

client = anthropic.Anthropic()
tool = get_anthropic_tool()

response = client.messages.create(
    model="claude-sonnet-4-20250514",
    max_tokens=1024,
    tools=[tool],
    messages=[{"role": "user", "content": "Analyze this sales data: [100, 120, 115, 500, 118]"}]
)

# Handle tool use in response
for block in response.content:
    if block.type == "tool_use" and block.name == "semantic_analysis":
        result = handle_tool_call(block.input)
        print(result)

LangChain

bash
pip install semantic-frame[langchain]
python
from semantic_frame.integrations.langchain import get_semantic_tool

tool = get_semantic_tool()
# Use as a LangChain BaseTool in your agent

CrewAI

bash
pip install semantic-frame[crewai]
python
from semantic_frame.integrations.crewai import get_crewai_tool

tool = get_crewai_tool()
# Use with CrewAI agents

MCP (Model Context Protocol)

bash
pip install semantic-frame[mcp]

Run the MCP server:

bash
mcp run semantic_frame.integrations.mcp:mcp

Exposes describe_data tool for MCP clients like:

  • ElizaOS: TypeScript-based agent framework
  • Claude Desktop: Anthropic's desktop app
  • Claude Code: Anthropic's CLI for Claude
  • Any MCP-compatible client

Claude Code

Add Semantic Frame as a native tool in Claude Code:

bash
# Install MCP dependencies
pip install semantic-frame[mcp]

# Add MCP server to Claude Code
claude mcp add semantic-frame -- uv run --project /path/to/semantic-frame mcp run /path/to/semantic-frame/semantic_frame/integrations/mcp.py

# Restart Claude Code, then verify connection
claude mcp list
# semantic-frame: ... - ✓ Connected

Once configured, ask Claude to analyze data and it will use the describe_data tool automatically.

Advanced Tool Use (Beta)

Semantic Frame supports Anthropic's Advanced Tool Use features for efficient tool orchestration in complex agent workflows.

Features

FeatureBenefitAPI
Input Examples+18% parameter accuracyIncluded by default
Tool Search1000+ tools without context bloatdefer_loading=True
Programmatic CallingBatch analysis via code executionallowed_callers=["code_execution"]

Quick Start (Advanced)

python
import anthropic
from semantic_frame.integrations.anthropic import get_advanced_tool, handle_tool_call

client = anthropic.Anthropic()
tool = get_advanced_tool()  # All advanced features enabled

response = client.beta.messages.create(
    betas=["advanced-tool-use-2025-11-20"],
    model="claude-sonnet-4-5-20250929",
    max_tokens=4096,
    tools=[
        {"type": "tool_search_tool_regex_20251119", "name": "tool_search"},
        {"type": "code_execution_20250825", "name": "code_execution"},
        tool,
    ],
    messages=[{"role": "user", "content": "Analyze all columns in this dataset..."}]
)

Configuration Options

python
from semantic_frame.integrations.anthropic import (
    get_anthropic_tool,          # Standard (includes examples)
    get_tool_for_discovery,      # For Tool Search
    get_tool_for_batch_processing,  # For code execution
    get_advanced_tool,           # All features enabled
)

MCP Batch Analysis

python
from semantic_frame.integrations.mcp import describe_batch

# Analyze multiple series in one call
result = describe_batch(
    datasets='{"cpu": [45, 47, 95, 44], "memory": [60, 61, 60, 61]}',
)

See docs/advanced-tool-use.md for complete documentation.

Use Cases

Crypto Trading

python
btc_prices = pd.Series(hourly_btc_prices)
insight = describe_series(btc_prices, context="BTC/USD Hourly")
# "The BTC/USD Hourly data shows a rapidly rising pattern with extreme variability.
#  Step up detected at index 142. 2 anomalies detected at indices 89, 203."

DevOps Monitoring

python
cpu_data = pd.Series(cpu_readings)
insight = describe_series(cpu_data, context="CPU Usage %")
# "The CPU Usage % data shows a flat/stationary pattern with stable variability
#  until index 850, where a critical anomaly was detected..."

Sales Analytics

python
sales = pd.Series(daily_sales)
insight = describe_series(sales, context="Daily Revenue")
# "The Daily Revenue data shows a steadily rising pattern with weak cyclic pattern
#  detected. Baseline: $12,450 (range: $8,200-$18,900)."

IoT Sensor Data

python
temps = pl.Series("temperature", sensor_readings)
insight = describe_series(temps, context="Machine Temperature (C)")
# "The Machine Temperature (C) data is expanding with extreme outliers.
#  3 anomalies detected at indices 142, 156, 161."

📈 Trading Module (v0.4.0)

Specialized semantic analysis for trading agents, portfolio managers, and financial applications.

Trading Tools

ToolDescription
describe_drawdownEquity curve drawdown analysis with severity
describe_trading_performanceWin rate, Sharpe, profit factor metrics
describe_rankingsMulti-agent/strategy comparison
describe_anomaliesEnhanced anomaly detection with PnL context
describe_windowsMulti-timeframe trend alignment
describe_regimeMarket regime detection (bull/bear/sideways)
describe_allocationPortfolio allocation suggestions ⚠️

Quick Examples

python
from semantic_frame.trading import (
    describe_trading_performance,
    describe_drawdown,
    describe_regime,
    describe_allocation,
)

# Trading Performance
pnl = [100, -50, 75, -25, 150, -30, 80]
result = describe_trading_performance(pnl, context="My Bot")
print(result.narrative)
# "My Bot shows good performance with 57.1% win rate. Profit factor: 2.53..."

# Drawdown Analysis
equity = [10000, 10500, 10200, 9800, 9500, 10100]
result = describe_drawdown(equity, context="Strategy")
print(result.narrative)
# "Strategy max drawdown: 9.5% (moderate). Currently recovering..."

# Market Regime
returns = [0.01, 0.015, 0.02, -0.01, 0.025, 0.018]  # Daily returns
result = describe_regime(returns, context="BTC")
print(result.narrative)
# "BTC is in a strong bullish regime. Conditions favor trend-following..."

# Portfolio Allocation (⚠️ Educational only, not financial advice)
assets = {"BTC": [40000, 42000, 44000], "ETH": [2500, 2650, 2800]}
result = describe_allocation(assets, method="risk_parity")
print(result.narrative)
# "Suggested allocation: BTC (55%), ETH (45%). Risk: high..."

MCP Integration

All trading tools are available via MCP:

bash
semantic-frame-mcp

Tools: describe_drawdown, describe_trading_performance, describe_rankings, describe_anomalies, describe_windows, describe_regime, describe_allocation

📖 Full Trading Documentation | Quick Reference

API Reference

describe_series(data, context=None, output="text")

Analyze a single data series.

Parameters:

  • data: Input data (NumPy array, Pandas Series, Polars Series, or list)
  • context: Optional label for the data (appears in narrative)
  • output: Format - "text" (string), "json" (dict), or "full" (SemanticResult)

Returns: Semantic description in requested format.

describe_dataframe(df, context=None)

Analyze all numeric columns in a DataFrame.

Parameters:

  • df: Pandas or Polars DataFrame
  • context: Optional prefix for column context labels

Returns: Dict mapping column names to SemanticResult objects.

SemanticResult

Full analysis result with:

  • narrative: Human-readable text description
  • trend: TrendState enum
  • volatility: VolatilityState enum
  • data_quality: DataQuality enum
  • anomaly_state: AnomalyState enum
  • anomalies: List of AnomalyInfo objects
  • seasonality: Optional SeasonalityState
  • distribution: Optional DistributionShape
  • step_change: Optional StructuralChange (STEP_UP, STEP_DOWN, NONE)
  • step_change_index: Optional int (index where step change occurred)
  • profile: SeriesProfile with statistics
  • compression_ratio: Token reduction ratio

Development

bash
# Clone and install
git clone https://github.com/yourusername/semantic-frame
cd semantic-frame
uv sync

# Run tests
uv run pytest

# Run with coverage
uv run pytest --cov=semantic_frame

License

MIT License - see LICENSE file.

常见问题

io.github.Anarkitty1/semantic-frame 是什么?

面向数值数据的高效语义压缩,显著节省 token,最高可实现 95% 以上 token reduction。

相关 Skills

网页构建器

by anthropics

Universal
热门

面向复杂 claude.ai HTML artifact 开发,快速初始化 React + Tailwind CSS + shadcn/ui 项目并打包为单文件 HTML,适合需要状态管理、路由或多组件交互的页面。

在 claude.ai 里做复杂网页 Artifact 很省心,多组件、状态和路由都能顺手搭起来,React、Tailwind 与 shadcn/ui 组合效率高、成品也更精致。

编码与调试
未扫描114.1k

前端设计

by anthropics

Universal
热门

面向组件、页面、海报和 Web 应用开发,按鲜明视觉方向生成可直接落地的前端代码与高质感 UI,适合做 landing page、Dashboard 或美化现有界面,避开千篇一律的 AI 审美。

想把页面做得既能上线又有设计感,就用前端设计:组件到整站都能产出,难得的是能避开千篇一律的 AI 味。

编码与调试
未扫描114.1k

网页应用测试

by anthropics

Universal
热门

用 Playwright 为本地 Web 应用编写自动化测试,支持启动开发服务器、校验前端交互、排查 UI 异常、抓取截图与浏览器日志,适合调试动态页面和回归验证。

借助 Playwright 一站式验证本地 Web 应用前端功能,调 UI 时还能同步查看日志和截图,定位问题更快。

编码与调试
未扫描114.1k

相关 MCP Server

GitHub

编辑精选

by GitHub

热门

GitHub 是 MCP 官方参考服务器,让 Claude 直接读写你的代码仓库和 Issues。

这个参考服务器解决了开发者想让 AI 安全访问 GitHub 数据的问题,适合需要自动化代码审查或 Issue 管理的团队。但注意它只是参考实现,生产环境得自己加固安全。

编码与调试
83.4k

by Context7

热门

Context7 是实时拉取最新文档和代码示例的智能助手,让你告别过时资料。

它能解决开发者查找文档时信息滞后的问题,特别适合快速上手新库或跟进更新。不过,依赖外部源可能导致偶尔的数据延迟,建议结合官方文档使用。

编码与调试
52.2k

by tldraw

热门

tldraw 是让 AI 助手直接在无限画布上绘图和协作的 MCP 服务器。

这解决了 AI 只能输出文本、无法视觉化协作的痛点——想象让 Claude 帮你画流程图或白板讨论。最适合需要快速原型设计或头脑风暴的开发者。不过,目前它只是个基础连接器,你得自己搭建画布应用才能发挥全部潜力。

编码与调试
46.3k

评论