com.falkordb/QueryWeaver
数据与存储by falkordb
面向 Text2SQL 的 MCP server,结合 graph schema 理解,将自然语言查询转换为可执行 SQL。
如果你常被写 SQL 卡住,QueryWeaver 能把自然语言直接转成可执行 SQL,还借助 graph schema 理解复杂表关系,结果更稳更准。
什么是 com.falkordb/QueryWeaver?
面向 Text2SQL 的 MCP server,结合 graph schema 理解,将自然语言查询转换为可执行 SQL。
README
REST API · MCP · Graph-powered
QueryWeaver is an open-source Text2SQL tool that converts plain-English questions into SQL using graph-powered schema understanding. It helps you ask databases natural-language questions and returns SQL and results.
</div>Get Started
Docker
💡 Recommended for evaluation purposes (Local Python or Node are not required)
docker run -p 5000:5000 -it falkordb/queryweaver
Launch: http://localhost:5000
Use an .env file (Recommended)
Create a local .env by copying .env.example and passing it to Docker. This is the simplest way to provide all required configuration:
cp .env.example .env
# edit .env to set your values, then:
docker run -p 5000:5000 --env-file .env falkordb/queryweaver
Alternative: Pass individual environment variables
If you prefer to pass variables on the command line, use -e flags (less convenient for many variables):
docker run -p 5000:5000 -it \
-e APP_ENV=production \
-e FASTAPI_SECRET_KEY=your_super_secret_key_here \
-e GOOGLE_CLIENT_ID=your_google_client_id \
-e GOOGLE_CLIENT_SECRET=your_google_client_secret \
-e GITHUB_CLIENT_ID=your_github_client_id \
-e GITHUB_CLIENT_SECRET=your_github_client_secret \
-e AZURE_API_KEY=your_azure_api_key \
falkordb/queryweaver
Note: QueryWeaver supports multiple AI providers. You can use
OPENAI_API_KEY,GEMINI_API_KEY,ANTHROPIC_API_KEY, orAZURE_API_KEY. See the AI/LLM configuration section for details.
For a full list of configuration options, consult
.env.example.
Memory TTL (optional)
QueryWeaver stores per-user conversation memory in FalkorDB. By default these graphs persist indefinitely. Set MEMORY_TTL_SECONDS to apply a Redis TTL (in seconds) so idle memory graphs are automatically cleaned up.
# Expire memory graphs after 1 week of inactivity
MEMORY_TTL_SECONDS=604800
The TTL is refreshed on every user interaction, so active users keep their memory.
MCP server: host or connect (optional)
QueryWeaver includes optional support for the Model Context Protocol (MCP). You can either have QueryWeaver expose an MCP-compatible HTTP surface (so other services can call QueryWeaver as an MCP server), or configure QueryWeaver to call an external MCP server for model/context services.
What QueryWeaver provides
-
The app registers MCP operations focused on Text2SQL flows:
list_databasesconnect_databasedatabase_schemaquery_database
-
To disable the built-in MCP endpoints set
DISABLE_MCP=truein your.envor environment (default: MCP enabled). -
Configuration
-
DISABLE_MCP— disable QueryWeaver's built-in MCP HTTP surface. Set totrueto disable. Default:false(MCP enabled).
Examples
Disable the built-in MCP when running with Docker:
docker run -p 5000:5000 -it --env DISABLE_MCP=true falkordb/queryweaver
Calling the built-in MCP endpoints (example)
- The MCP surface is exposed as HTTP endpoints.
Server Configuration
Below is a minimal example mcp.json client configuration that targets a local QueryWeaver instance exposing the MCP HTTP surface at /mcp.
{
"servers": {
"queryweaver": {
"type": "http",
"url": "http://127.0.0.1:5000/mcp",
"headers": {
"Authorization": "Bearer your_token_here"
}
}
},
"inputs": []
}
REST API
API Documentation
Swagger UI: https://app.queryweaver.ai/docs
OpenAPI JSON: https://app.queryweaver.ai/openapi.json
Overview
QueryWeaver exposes a small REST API for managing graphs (database schemas) and running Text2SQL queries. All endpoints that modify or access user-scoped data require authentication via a bearer token. In the browser the app uses session cookies and OAuth flows; for CLI and scripts you can use an API token (see tokens routes or the web UI to create one).
Core endpoints
- GET /graphs — list available graphs for the authenticated user
- GET /graphs/{graph_id}/data — return nodes/links (tables, columns, foreign keys) for the graph
- POST /graphs — upload or create a graph (JSON payload or file upload)
- POST /graphs/{graph_id} — run a Text2SQL chat query against the named graph (streaming response)
Authentication
- Add an Authorization header:
Authorization: Bearer <API_TOKEN>
Examples
- List graphs (GET)
curl example:
curl -s -H "Authorization: Bearer $TOKEN" \
https://app.queryweaver.ai/graphs
Python example:
import requests
resp = requests.get('https://app.queryweaver.ai/graphs', headers={'Authorization': f'Bearer {TOKEN}'})
print(resp.json())
- Get graph schema (GET)
curl example:
curl -s -H "Authorization: Bearer $TOKEN" \
https://app.queryweaver.ai/graphs/my_database/data
Python example:
resp = requests.get('https://app.queryweaver.ai/graphs/my_database/data', headers={'Authorization': f'Bearer {TOKEN}'})
print(resp.json())
- Load a graph (POST) — JSON payload
curl -H "Authorization: Bearer $TOKEN" -H "Content-Type: application/json" \
-d '{"database": "my_database", "tables": [...]}' \
https://app.queryweaver.ai/graphs
Or upload a file (multipart/form-data):
curl -H "Authorization: Bearer $TOKEN" -F "file=@schema.json" \
https://app.queryweaver.ai/graphs
- Query a graph (POST) — run a chat-based Text2SQL request
The POST /graphs/{graph_id} endpoint accepts a JSON body with at least a chat field (an array of messages). The endpoint streams processing steps and the final SQL back as server-sent-message chunks delimited by a special boundary used by the frontend. For simple scripting you can call it and read the final JSON object from the streamed messages.
Example payload:
{
"chat": ["How many users signed up last month?"],
"result": [],
"instructions": "Prefer PostgreSQL compatible SQL"
}
curl example (simple, collects whole response):
curl -s -H "Authorization: Bearer $TOKEN" -H "Content-Type: application/json" \
-d '{"chat": ["Count orders last week"]}' \
https://app.queryweaver.ai/graphs/my_database
Python example (stream-aware):
import requests
import json
url = 'https://app.queryweaver.ai/graphs/my_database'
headers = {'Authorization': f'Bearer {TOKEN}', 'Content-Type': 'application/json'}
with requests.post(url, headers=headers, json={"chat": ["Count orders last week"]}, stream=True) as r:
# The server yields JSON objects delimited by a message boundary string
boundary = '|||FALKORDB_MESSAGE_BOUNDARY|||'
buffer = ''
for chunk in r.iter_content(decode_unicode=True, chunk_size=1024):
buffer += chunk
while boundary in buffer:
part, buffer = buffer.split(boundary, 1)
if not part.strip():
continue
obj = json.loads(part)
print('STREAM:', obj)
Notes & tips
- Graph IDs are namespaced per-user. When calling the API directly use the plain graph id (the server will namespace by the authenticated user). For uploaded files the
databasefield determines the saved graph id. - The streaming response includes intermediate reasoning steps, follow-up questions (if the query is ambiguous or off-topic), and the final SQL. The frontend expects the boundary string
|||FALKORDB_MESSAGE_BOUNDARY|||between messages. - For destructive SQL (INSERT/UPDATE/DELETE etc) the service will include a confirmation step in the stream; the frontend handles this flow. If you automate destructive operations, ensure you handle confirmation properly (see the
ConfirmRequestmodel in the code).
Development
Follow these steps to run and develop QueryWeaver from source.
Prerequisites
- Python 3.12+
- uv (Python package manager)
- A FalkorDB instance (local or remote)
- Node.js and npm (for the React frontend)
Install and configure
Quickstart (recommended for development):
# Clone the repo
git clone https://github.com/FalkorDB/QueryWeaver.git
cd QueryWeaver
# Install dependencies (backend + frontend) and start the dev server
make install
make run-dev
If you prefer to set up manually or need a custom environment, use uv:
# Install Python (backend) and frontend dependencies
uv sync
# Create a local environment file
cp .env.example .env
# Edit .env with your values (set APP_ENV=development for local development)
Run the app locally
uv run uvicorn api.index:app --host 0.0.0.0 --port 5000 --reload
The server will be available at http://localhost:5000
Alternatively, the repository provides Make targets for running the app:
make run-dev # development server (reload, debug-friendly)
make run-prod # production mode (ensure frontend build if needed)
Frontend build (when needed)
The frontend is a modern React + Vite app in app/. Build before production runs or after frontend changes:
make install # installs backend and frontend deps
make build-prod # builds the frontend into app/dist/
# or manually
cd app
npm ci
npm run build
OAuth configuration
QueryWeaver supports Google and GitHub OAuth. Create OAuth credentials for each provider and paste the client IDs/secrets into your .env file.
- Google: set authorized origin and callback
http://localhost:5000/login/google/authorized - GitHub: set homepage and callback
http://localhost:5000/login/github/authorized
Environment-specific OAuth settings
For production/staging deployments, set APP_ENV=production or APP_ENV=staging in your environment to enable secure session cookies (HTTPS-only). This prevents OAuth CSRF state mismatch errors.
# For production/staging (enables HTTPS-only session cookies)
APP_ENV=production
# For development (allows HTTP session cookies)
APP_ENV=development
Important: If you're getting "mismatching_state: CSRF Warning!" errors on staging/production, ensure APP_ENV is set to production or staging to enable secure session handling.
AI/LLM configuration
QueryWeaver supports multiple AI providers. Set one API key and QueryWeaver auto-detects which provider to use.
Priority order: Ollama > OpenAI > Gemini > Anthropic > Cohere > Azure (default)
| Provider | API Key | Default Models |
|---|---|---|
| Ollama | OLLAMA_MODEL | ollama/<your-model>, ollama/nomic-embed-text |
| OpenAI | OPENAI_API_KEY | openai/gpt-4.1, openai/text-embedding-ada-002 |
| Google Gemini | GEMINI_API_KEY | gemini/gemini-3-pro-preview, gemini/gemini-embedding-001 |
| Anthropic | ANTHROPIC_API_KEY | anthropic/claude-sonnet-4-5-20250929, voyage/voyage-3* |
| Cohere | COHERE_API_KEY | cohere/command-a-03-2025, cohere/embed-v4.0 |
| Azure OpenAI | AZURE_API_KEY | azure/gpt-4.1, azure/text-embedding-ada-002 |
* Anthropic has no native embeddings. You must set VOYAGE_API_KEY or EMBEDDING_MODEL for embeddings, otherwise startup will fail with an error.
Optional: Override default models
COMPLETION_MODEL=gemini/gemini-3-pro-preview
EMBEDDING_MODEL=gemini/gemini-embedding-001
Both must match your API key's provider.
Docker examples with AI configuration
Using OpenAI:
docker run -p 5000:5000 -it \
-e FASTAPI_SECRET_KEY=your_secret_key \
-e OPENAI_API_KEY=your_openai_api_key \
falkordb/queryweaver
Using Google Gemini:
docker run -p 5000:5000 -it \
-e FASTAPI_SECRET_KEY=your_secret_key \
-e GEMINI_API_KEY=your_gemini_api_key \
falkordb/queryweaver
Using Anthropic:
docker run -p 5000:5000 -it \
-e FASTAPI_SECRET_KEY=your_secret_key \
-e ANTHROPIC_API_KEY=your_anthropic_api_key \
falkordb/queryweaver
Using Azure OpenAI:
docker run -p 5000:5000 -it \
-e FASTAPI_SECRET_KEY=your_secret_key \
-e AZURE_API_KEY=your_azure_api_key \
-e AZURE_API_BASE=https://your-resource.openai.azure.com/ \
-e AZURE_API_VERSION=2024-12-01-preview \
falkordb/queryweaver
Testing
Quick note: many tests require FalkorDB to be available. Use the included helper to run a test DB in Docker if needed.
Prerequisites
- Install dev dependencies:
uv sync - Start FalkorDB (see
make docker-falkordb) - Install Playwright browsers:
uv run playwright install
Quick commands
Recommended: prepare the development/test environment using the Make helper (installs dependencies and Playwright browsers):
# Prepare development/test environment (installs deps and Playwright browsers)
make setup-dev
Alternatively, you can run the E2E-specific setup script and then run tests manually:
# Prepare E2E test environment (installs browsers and other setup)
./setup_e2e_tests.sh
# Run all tests
make test
# Run unit tests only (faster)
make test-unit
# Run E2E tests (headless)
make test-e2e
# Run E2E tests with a visible browser for debugging
make test-e2e-headed
Test types
- Unit tests: focus on individual modules and utilities. Run with
make test-unitoruv run python -m pytest tests/ -k "not e2e". - End-to-end (E2E) tests: run via Playwright and exercise UI flows, OAuth, file uploads, schema processing, chat queries, and API endpoints. Use
make test-e2e.
See tests/e2e/README.md for full E2E test instructions.
CI/CD
GitHub Actions run unit and E2E tests on pushes and pull requests. Failures capture screenshots and artifacts for debugging.
Troubleshooting
- FalkorDB connection issues: start the DB helper
make docker-falkordbor check network/host settings. - Playwright/browser failures: install browsers with
uv run playwright installand ensure system deps are present. - Missing environment variables: copy
.env.exampleand fill required values. - OAuth "mismatching_state: CSRF Warning!" errors: Set
APP_ENV=production(orstaging) in your environment for HTTPS deployments, orAPP_ENV=developmentfor HTTP development environments. This ensures session cookies are configured correctly for your deployment type.
Project layout (high level)
api/– FastAPI backendapp/– React + Vite frontendtests/– unit and E2E tests
License
Licensed under the GNU Affero General Public License (AGPL). See LICENSE.
Copyright FalkorDB Ltd. 2025
常见问题
com.falkordb/QueryWeaver 是什么?
面向 Text2SQL 的 MCP server,结合 graph schema 理解,将自然语言查询转换为可执行 SQL。
相关 Skills
迁移架构师
by alirezarezvani
为数据库、API 与基础设施迁移制定分阶段零停机方案,提前校验兼容性与风险,生成回滚策略、验证关卡和时间线,适合复杂系统平滑切换。
✎ 做数据库与存储迁移时,用它统一梳理表结构和数据搬迁流程,架构视角更完整,复杂迁移也更稳。
数据库建模
by alirezarezvani
把需求梳理成关系型数据库表结构,自动生成迁移脚本、TypeScript/Python 类型、种子数据、RLS 策略和索引方案,适合多租户、审计追踪、软删除等后端建模与 Schema 评审场景。
✎ 把数据库结构设计、ER图梳理和SQL建模放到一处,复杂业务也能快速统一数据模式,少走不少返工弯路。
资深数据工程师
by alirezarezvani
聚焦生产级数据工程,覆盖 ETL/ELT、批处理与流式管道、数据建模、Airflow/dbt/Spark 优化和数据质量治理,适合设计数据架构、搭建现代数据栈与排查性能问题。
✎ 复杂数据管道、ETL/ELT 和治理难题交给它,凭 Spark、Airflow、dbt 等现代数据栈经验,能更稳地搭起可扩展的数据基础设施。
相关 MCP Server
PostgreSQL 数据库
编辑精选by Anthropic
PostgreSQL 是让 Claude 直接查询和管理你的数据库的 MCP 服务器。
✎ 这个服务器解决了开发者需要手动编写 SQL 查询的痛点,特别适合数据分析师或后端开发者快速探索数据库结构。不过,由于是参考实现,生产环境使用前务必评估安全风险,别指望它能处理复杂事务。
SQLite 数据库
编辑精选by Anthropic
SQLite 是让 AI 直接查询本地数据库进行数据分析的 MCP 服务器。
✎ 这个服务器解决了 AI 无法直接访问 SQLite 数据库的问题,适合需要快速分析本地数据集的开发者。不过,作为参考实现,它可能缺乏生产级的安全特性,建议在受控环境中使用。
Firecrawl 智能爬虫
编辑精选by Firecrawl
Firecrawl 是让 AI 直接抓取网页并提取结构化数据的 MCP 服务器。
✎ 它解决了手动写爬虫的麻烦,让 Claude 能直接访问动态网页内容。最适合需要实时数据的研究者或开发者,比如监控竞品价格或抓取新闻。但要注意,它依赖第三方 API,可能涉及隐私和成本问题。