Databricks MCP Server
平台与服务by pramodbhatofficial
提供 263 个 Databricks 工具,覆盖 Unity Catalog、SQL、Compute、Jobs、Serving 等核心能力。
什么是 Databricks MCP Server?
提供 263 个 Databricks 工具,覆盖 Unity Catalog、SQL、Compute、Jobs、Serving 等核心能力。
README
Databricks MCP Server
<!-- mcp-name: io.github.pramodbhatofficial/databricks-sdk-mcp -->A comprehensive Model Context Protocol (MCP) server for Databricks, built on the official Databricks Python SDK.
Provides 263 tools and 8 prompt templates across 28 service domains, giving AI assistants full access to the Databricks platform.
Features
- SDK-first: Uses
databricks-sdkfor type safety and automatic API freshness - Comprehensive: Covers Unity Catalog, SQL, Compute, Jobs, Pipelines, Serving, Vector Search, Apps, Lakebase, Dashboards, Genie, Secrets, IAM, Connections, Experiments, and Delta Sharing
- Zero custom auth: Delegates authentication entirely to the SDK (PAT, OAuth, Azure AD, service principal -- all automatic)
- Selective loading: Include/exclude tool modules via environment variables
- MCP Resources: Read-only workspace context (URL, current user, auth type)
Quick Start
Installation
pip install databricks-sdk-mcp
Or run with Docker:
docker run -i -e DATABRICKS_HOST=... -e DATABRICKS_TOKEN=... databricks-mcp
Or install from source:
git clone https://github.com/pramodbhatofficial/databricks-mcp-server.git
cd databricks-mcp-server
pip install -e ".[dev]"
Authentication
Authentication is handled by the Databricks SDK. Set one of:
Personal Access Token (simplest):
export DATABRICKS_HOST=https://your-workspace.databricks.com
export DATABRICKS_TOKEN=dapi...
OAuth (M2M):
export DATABRICKS_HOST=https://your-workspace.databricks.com
export DATABRICKS_CLIENT_ID=...
export DATABRICKS_CLIENT_SECRET=...
Other methods: Azure AD, Databricks CLI profile, Azure Managed Identity -- all auto-detected by the SDK.
Running
databricks-mcp
This starts the MCP server using stdio transport.
Integrations
Claude Code (Terminal)
Add to ~/.claude/settings.json or your project's .claude/settings.json:
{
"mcpServers": {
"databricks": {
"command": "databricks-mcp",
"env": {
"DATABRICKS_HOST": "https://your-workspace.databricks.com",
"DATABRICKS_TOKEN": "dapi..."
}
}
}
}
Then restart Claude Code. Verify with /mcp to see the registered tools.
Claude Desktop
Add to your Claude Desktop config file:
- macOS:
~/Library/Application Support/Claude/claude_desktop_config.json - Windows:
%APPDATA%\Claude\claude_desktop_config.json
{
"mcpServers": {
"databricks": {
"command": "databricks-mcp",
"env": {
"DATABRICKS_HOST": "https://your-workspace.databricks.com",
"DATABRICKS_TOKEN": "dapi..."
}
}
}
}
Restart Claude Desktop. The Databricks tools will appear in the tool picker.
Cursor
Add to .cursor/mcp.json in your project root (or ~/.cursor/mcp.json for global):
{
"mcpServers": {
"databricks": {
"command": "databricks-mcp",
"env": {
"DATABRICKS_HOST": "https://your-workspace.databricks.com",
"DATABRICKS_TOKEN": "dapi..."
}
}
}
}
Open Cursor Settings > MCP to verify the server is connected.
Windsurf
Add to ~/.codeium/windsurf/mcp_config.json:
{
"mcpServers": {
"databricks": {
"command": "databricks-mcp",
"env": {
"DATABRICKS_HOST": "https://your-workspace.databricks.com",
"DATABRICKS_TOKEN": "dapi..."
}
}
}
}
VS Code (Copilot)
Add to .vscode/mcp.json in your project:
{
"servers": {
"databricks": {
"command": "databricks-mcp",
"env": {
"DATABRICKS_HOST": "https://your-workspace.databricks.com",
"DATABRICKS_TOKEN": "dapi..."
}
}
}
}
Zed
Add to Zed's settings (~/.config/zed/settings.json):
{
"context_servers": {
"databricks": {
"command": {
"path": "databricks-mcp",
"env": {
"DATABRICKS_HOST": "https://your-workspace.databricks.com",
"DATABRICKS_TOKEN": "dapi..."
}
}
}
}
}
Any MCP Client (Generic stdio)
The server uses stdio transport. Connect from any MCP-compatible client:
# Set auth env vars
export DATABRICKS_HOST=https://your-workspace.databricks.com
export DATABRICKS_TOKEN=dapi...
# Start the server (communicates via stdin/stdout)
databricks-mcp
Tip: Load Only What You Need
If your MCP client struggles with many tools, use selective loading to reduce the tool count:
{
"mcpServers": {
"databricks": {
"command": "databricks-mcp",
"env": {
"DATABRICKS_HOST": "https://your-workspace.databricks.com",
"DATABRICKS_TOKEN": "dapi...",
"DATABRICKS_MCP_TOOLS_INCLUDE": "unity_catalog,sql,compute,jobs"
}
}
}
}
Tool Modules
| Module | Tools | Description |
|---|---|---|
unity_catalog | 23 | Catalogs, schemas, tables, volumes, functions, registered models |
sql | 14 | Warehouses, SQL execution, queries, alerts, history |
workspace | 10 | Notebooks, files, repos |
compute | 18 | Clusters, instance pools, policies, node types, Spark versions |
jobs | 13 | Jobs, runs, tasks, repair, cancel all |
pipelines | 8 | DLT / Lakeflow pipelines |
serving | 10 | Serving endpoints, model versions, OpenAPI |
vector_search | 10 | Vector search endpoints, indexes, sync |
apps | 10 | Databricks Apps lifecycle |
database | 10 | Lakebase PostgreSQL instances |
dashboards | 9 | Lakeview AI/BI dashboards, published views |
genie | 5 | Genie AI/BI conversations |
secrets | 8 | Secret scopes and secrets |
iam | 16 | Users, groups, service principals, permissions, current user |
connections | 5 | External connections |
experiments | 14 | MLflow experiments, runs, artifacts, metrics, params |
sharing | 11 | Delta Sharing shares, recipients, providers |
files | 12 | DBFS and UC Volumes file operations |
grants | 3 | Unity Catalog permission grants (GRANT/REVOKE) |
storage | 10 | Storage credentials and external locations |
metastores | 8 | Unity Catalog metastore management |
online_tables | 3 | Online tables for low-latency serving |
global_init_scripts | 5 | Workspace-wide init scripts |
tokens | 5 | Personal access token management |
git_credentials | 5 | Git credential management for repos |
quality_monitors | 8 | Data quality monitoring and refreshes |
command_execution | 4 | Interactive command execution on clusters |
workflows | 5 | Composite multi-step operations (workspace status, schema setup, query preview) |
Selective Tool Loading
With 263 tools, it's recommended to load only the modules you need. This improves agent performance and tool selection accuracy.
Role-Based Presets (Recommended)
Pick a preset that matches your role:
| Preset | Modules | Tools | Config |
|---|---|---|---|
| Data Engineer | unity_catalog, sql, compute, jobs, pipelines, files, quality_monitors | ~100 | DATABRICKS_MCP_TOOLS_INCLUDE=unity_catalog,sql,compute,jobs,pipelines,files,quality_monitors |
| ML Engineer | serving, vector_search, experiments, compute, unity_catalog, online_tables, files | ~98 | DATABRICKS_MCP_TOOLS_INCLUDE=serving,vector_search,experiments,compute,unity_catalog,online_tables,files |
| Platform Admin | iam, secrets, tokens, metastores, compute, global_init_scripts, grants, storage | ~85 | DATABRICKS_MCP_TOOLS_INCLUDE=iam,secrets,tokens,metastores,compute,global_init_scripts,grants,storage |
| App Developer | apps, database, sql, files, serving, secrets | ~64 | DATABRICKS_MCP_TOOLS_INCLUDE=apps,database,sql,files,serving,secrets |
| Data Analyst | sql, unity_catalog, dashboards, genie, workspace | ~61 | DATABRICKS_MCP_TOOLS_INCLUDE=sql,unity_catalog,dashboards,genie,workspace |
| Minimal | sql, unity_catalog | ~37 | DATABRICKS_MCP_TOOLS_INCLUDE=sql,unity_catalog |
Example using a preset in Claude Code:
{
"mcpServers": {
"databricks": {
"command": "databricks-mcp",
"env": {
"DATABRICKS_HOST": "https://your-workspace.databricks.com",
"DATABRICKS_TOKEN": "dapi...",
"DATABRICKS_MCP_TOOLS_INCLUDE": "unity_catalog,sql,compute,jobs,pipelines,files,quality_monitors"
}
}
}
}
Custom Filtering
# Only include specific modules
export DATABRICKS_MCP_TOOLS_INCLUDE=unity_catalog,sql,serving
# Exclude specific modules (cannot combine with INCLUDE)
export DATABRICKS_MCP_TOOLS_EXCLUDE=iam,sharing,experiments
If INCLUDE is set, only those modules load. If EXCLUDE is set, everything except those modules loads. INCLUDE takes precedence if both are set.
Tool Discovery (For AI Agents)
The server includes built-in tool discovery to help AI agents find the right tools:
MCP Resources
| URI | Description |
|---|---|
databricks://workspace/info | Workspace URL, current user, auth type |
databricks://tools/guide | Tool catalog with module descriptions, use cases, and role presets |
Agents can read databricks://tools/guide at connection time to understand what's available.
Discovery Tool
The databricks_tool_guide tool helps agents find the right tools during a conversation:
# Find tools for a specific task
databricks_tool_guide(task="run a SQL query")
databricks_tool_guide(task="deploy an ML model")
databricks_tool_guide(task="create a user")
# Get role-based recommendations
databricks_tool_guide(role="data_engineer")
databricks_tool_guide(role="ml_engineer")
This returns matching modules with descriptions and usage hints, so the agent knows exactly which databricks_* tools to call.
MCP Prompts (Guided Workflows)
The server includes 8 prompt templates that guide AI agents through multi-step Databricks workflows:
| Prompt | Description |
|---|---|
explore_data_catalog | Browse Unity Catalog structure (catalogs → schemas → tables) |
query_data | Find a warehouse, execute SQL, and format results |
debug_failing_job | Investigate a failing job: status, logs, error analysis |
setup_ml_experiment | Create an MLflow experiment and configure tracking |
deploy_model | Deploy a model to a serving endpoint |
setup_data_pipeline | Create a DLT pipeline with scheduling |
workspace_health_check | Audit clusters, warehouses, jobs, and endpoints |
manage_permissions | Review and update permissions on workspace objects |
Prompts appear automatically in MCP clients that support them (e.g., Claude Desktop's prompt picker).
Docker
Run the MCP server in a container:
# Build
docker build -t databricks-mcp .
# Run with stdio
docker run -i \
-e DATABRICKS_HOST=https://your-workspace.databricks.com \
-e DATABRICKS_TOKEN=dapi... \
databricks-mcp
# Run with SSE transport
docker run -p 8080:8080 \
-e DATABRICKS_HOST=https://your-workspace.databricks.com \
-e DATABRICKS_TOKEN=dapi... \
databricks-mcp --transport sse --port 8080
# Run with selective modules
docker run -i \
-e DATABRICKS_HOST=https://your-workspace.databricks.com \
-e DATABRICKS_TOKEN=dapi... \
-e DATABRICKS_MCP_TOOLS_INCLUDE=sql,unity_catalog \
databricks-mcp
SSE Transport (Remote Server)
The server supports SSE transport for remote connections:
# Start as SSE server
databricks-mcp --transport sse --port 8080
# Custom host/port
databricks-mcp --transport sse --host 127.0.0.1 --port 3000
Connect from any MCP client that supports SSE:
{
"mcpServers": {
"databricks": {
"url": "http://localhost:8080/sse"
}
}
}
Development
# Install with dev dependencies
pip install -e ".[dev]"
# Lint
ruff check databricks_mcp/
# Test
pytest tests/ -v
Author
Pramod Bhat
License
Apache 2.0 -- see LICENSE.
常见问题
Databricks MCP Server 是什么?
提供 263 个 Databricks 工具,覆盖 Unity Catalog、SQL、Compute、Jobs、Serving 等核心能力。
相关 Skills
MCP构建
by anthropics
聚焦高质量 MCP Server 开发,覆盖协议研究、工具设计、错误处理与传输选型,适合用 FastMCP 或 MCP SDK 对接外部 API、封装服务能力。
✎ 想让 LLM 稳定调用外部 API,就用 MCP构建:从 Python 到 Node 都有成熟指引,帮你更快做出高质量 MCP 服务器。
Slack动图
by anthropics
面向Slack的动图制作Skill,内置emoji/消息GIF的尺寸、帧率和色彩约束、校验与优化流程,适合把创意或上传图片快速做成可直接发送的Slack动画。
✎ 帮你快速做出适配 Slack 的动图,内置约束规则和校验工具,少踩上传与播放坑,做表情包和演示都更省心。
MCP服务构建器
by alirezarezvani
从 OpenAPI 一键生成 Python/TypeScript MCP server 脚手架,并校验 tool schema、命名规范与版本兼容性,适合把现有 REST API 快速发布成可生产演进的 MCP 服务。
✎ 帮你快速搭建 MCP 服务与后端 API,脚手架完善、扩展顺手,尤其适合想高效验证服务能力的开发者。
相关 MCP Server
Slack 消息
编辑精选by Anthropic
Slack 是让 AI 助手直接读写你的 Slack 频道和消息的 MCP 服务器。
✎ 这个服务器解决了团队协作中需要 AI 实时获取 Slack 信息的痛点,特别适合开发团队让 Claude 帮忙汇总频道讨论或发送通知。不过,它目前只是参考实现,文档有限,不建议在生产环境直接使用——更适合开发者学习 MCP 如何集成第三方服务。
by netdata
io.github.netdata/mcp-server 是让 AI 助手实时监控服务器指标和日志的 MCP 服务器。
✎ 这个工具解决了运维人员需要手动检查系统状态的痛点,最适合 DevOps 团队让 Claude 自动分析性能数据。不过,它依赖 NetData 的现有部署,如果你没用过这个监控平台,得先花时间配置。
by d4vinci
Scrapling MCP Server 是专为现代网页设计的智能爬虫工具,支持绕过 Cloudflare 等反爬机制。
✎ 这个工具解决了爬取动态网页和反爬网站时的头疼问题,特别适合需要批量采集电商价格或新闻数据的开发者。不过,它依赖外部浏览器引擎,资源消耗较大,不适合轻量级任务。