ai.smithery/IlyaGusev-academia_mcp

AI 与智能体

by ilyagusev

搜索 arXiv 与 ACL Anthology,获取 citations 和 references,并浏览网页资料以加速学术研究。

把 arXiv 和 ACL Anthology 检索、引文追踪整合到一起,查论文和参考文献更省时间,做 NLP 学术调研尤其顺手。

什么是 ai.smithery/IlyaGusev-academia_mcp

搜索 arXiv 与 ACL Anthology,获取 citations 和 references,并浏览网页资料以加速学术研究。

README

Academia MCP

PyPI CI License smithery badge Verified on MseeP

MCP server with tools to search, fetch, analyze, and report on scientific papers and datasets.

Features

  • ArXiv search and download
  • ACL Anthology search
  • Hugging Face datasets search
  • Semantic Scholar citations and references
  • Web search via Exa, Brave, or Tavily
  • Web page crawler, LaTeX compilation, PDF reading
  • Optional LLM-powered tools for document QA and research proposal workflows

Requirements

  • Python 3.12+

Install

  • Using pip (end users):
bash
pip3 install academia-mcp
  • For development (uv + Makefile):
bash
uv venv .venv
make install

Quickstart

  • Run over HTTP (default transport):
bash
python -m academia_mcp --transport streamable-http
# OR
uv run -m academia_mcp --transport streamable-http
  • Run over stdio (for local MCP clients like Claude Desktop):
bash
python -m academia_mcp --transport stdio
# OR
uv run -m academia_mcp --transport stdio

Notes:

  • Transports: stdio, sse, streamable-http.
  • host/port are used for HTTP transports; ignored for stdio. Default port is 5056 (or PORT).

Authentication

Academia MCP supports optional token-based authentication for HTTP transports (streamable-http and sse). Authentication is disabled by default to maintain backward compatibility.

Enabling Authentication

Set the ENABLE_AUTH environment variable to true:

bash
export ENABLE_AUTH=true
export TOKENS_FILE=/path/to/tokens.json  # Optional, defaults to ./tokens.json

Managing Tokens

Issue a new token:

bash
academia_mcp auth issue-token --client-id=my-client --description="Production API client"

# Issue token with 30-day expiration
academia_mcp auth issue-token --client-id=test-client --expires-days=30

# Issue token with custom scopes
academia_mcp auth issue-token --client-id=admin --scopes="read,write,admin"

List active tokens:

bash
academia_mcp auth list-tokens

Revoke a token:

bash
academia_mcp auth revoke-token mcp_a1b2c3d4e5f6...

Using Tokens

Include the token in the Authorization header with the Bearer scheme or as a query parameter apiKey.

Security Notes:

  • Tokens are displayed only once during issuance. Store them securely.
  • Use HTTPS in production to protect tokens in transit.
  • The tokens.json file is automatically created with restrictive permissions (mode 600).
  • Tokens are stored in plaintext (standard practice for bearer tokens) - protect the tokens file.

Claude Desktop config

json
{
  "mcpServers": {
    "academia": {
      "command": "python3",
      "args": [
        "-m",
        "academia_mcp",
        "--transport",
        "stdio"
      ]
    }
  }
}

Available tools (one-liners)

  • arxiv_search: Query arXiv with field-specific queries and filters.
  • arxiv_download: Fetch a paper by ID and convert to structured text (HTML/PDF modes).
  • anthology_search: Search ACL Anthology with fielded queries and optional date filtering.
  • hf_datasets_search: Find Hugging Face datasets with filters and sorting.
  • s2_get_citations: List papers citing a given arXiv paper (Semantic Scholar Graph).
  • s2_get_references: List papers referenced by a given arXiv paper.
  • visit_webpage: Fetch and normalize a web page.
  • web_search: Unified search wrapper; available when at least one of Exa/Brave/Tavily keys is set.
  • exa_web_search, brave_web_search, tavily_web_search: Provider-specific search.
  • get_latex_templates_list, get_latex_template: Enumerate and fetch built-in LaTeX templates.
  • compile_latex: Compile LaTeX to PDF in WORKSPACE_DIR.
  • read_pdf: Extract text per page from a PDF.
  • download_pdf_paper, review_pdf_paper: Download and optionally review PDFs (requires LLM + workspace).
  • document_qa: Answer questions over provided document chunks (requires LLM).
  • extract_bitflip_info, generate_research_proposals, score_research_proposals: Research proposal helpers (requires LLM).

Availability notes:

  • Set WORKSPACE_DIR to enable compile_latex, read_pdf, download_pdf_paper, and review_pdf_paper.
  • Set OPENROUTER_API_KEY to enable LLM tools (document_qa, review_pdf_paper, and bitflip tools).
  • Set one or more of EXA_API_KEY, BRAVE_API_KEY, TAVILY_API_KEY to enable web_search and provider tools.

Environment variables

Set as needed, depending on which tools you use:

  • OPENROUTER_API_KEY: required for LLM-related tools.
  • BASE_URL: override OpenRouter base URL.
  • DOCUMENT_QA_MODEL_NAME: override default model for document_qa.
  • BITFLIP_MODEL_NAME: override default model for bitflip tools.
  • TAVILY_API_KEY: enables Tavily in web_search.
  • EXA_API_KEY: enables Exa in web_search and visit_webpage.
  • BRAVE_API_KEY: enables Brave in web_search.
  • WORKSPACE_DIR: directory for generated files (PDFs, temp artifacts).
  • PORT: HTTP port (default 5056).

You can put these in a .env file in the project root.

Docker

Build the image:

bash
docker build -t academia_mcp .

Run the server (HTTP):

bash
docker run --rm -p 5056:5056 \
  -e PORT=5056 \
  -e OPENROUTER_API_KEY=your_key_here \
  -e WORKSPACE_DIR=/workspace \
  -v "$PWD/workdir:/workspace" \
  academia_mcp

Or use existing image: phoenix120/academia_mcp

Examples

Makefile targets

  • make install: install the package in editable mode with uv
  • make validate: run black, flake8, and mypy (strict)
  • make test: run the test suite with pytest
  • make publish: build and publish using uv

LaTeX/PDF requirements

Only needed for LaTeX/PDF tools. Ensure a LaTeX distribution is installed and pdflatex is on PATH, as well as latexmk. On Debian/Ubuntu:

bash
sudo apt install texlive-latex-base texlive-fonts-recommended texlive-latex-extra texlive-science latexmk

常见问题

ai.smithery/IlyaGusev-academia_mcp 是什么?

搜索 arXiv 与 ACL Anthology,获取 citations 和 references,并浏览网页资料以加速学术研究。

相关 Skills

Claude接口

by anthropics

Universal
热门

面向接入 Claude API、Anthropic SDK 或 Agent SDK 的开发场景,自动识别项目语言并给出对应示例与默认配置,快速搭建 LLM 应用。

想把Claude能力接进应用或智能体,用claude-api上手快、兼容Anthropic与Agent SDK,集成路径清晰又省心

AI 与智能体
未扫描109.6k

提示工程专家

by alirezarezvani

Universal
热门

覆盖Prompt优化、Few-shot设计、结构化输出、RAG评测与Agent工作流编排,适合分析token成本、评估LLM输出质量,并搭建可落地的AI智能体系统。

把提示优化、LLM评测到RAG与智能体设计串成一套方法,适合想系统提升AI开发效率的人。

AI 与智能体
未扫描9.0k

智能体流程设计

by alirezarezvani

Universal
热门

面向生产级多 Agent 编排,梳理顺序、并行、分层、事件驱动、共识五种工作流设计,覆盖 handoff、状态管理、容错重试、上下文预算与成本优化,适合搭建复杂 AI 协作系统。

帮你把多智能体流程设计、编排和自动化统一起来,复杂工作流也能更稳地落地,适合追求强控制力的团队。

AI 与智能体
未扫描9.0k

相关 MCP Server

顺序思维

编辑精选

by Anthropic

热门

Sequential Thinking 是让 AI 通过动态思维链解决复杂问题的参考服务器。

这个服务器展示了如何让 Claude 像人类一样逐步推理,适合开发者学习 MCP 的思维链实现。但注意它只是个参考示例,别指望直接用在生产环境里。

AI 与智能体
82.9k

知识图谱记忆

编辑精选

by Anthropic

热门

Memory 是一个基于本地知识图谱的持久化记忆系统,让 AI 记住长期上下文。

帮 AI 和智能体补上“记不住”的短板,用本地知识图谱沉淀长期上下文,连续对话更聪明,数据也更可控。

AI 与智能体
82.9k

PraisonAI

编辑精选

by mervinpraison

热门

PraisonAI 是一个支持自反思和多 LLM 的低代码 AI 智能体框架。

如果你需要快速搭建一个能 24/7 运行的 AI 智能体团队来处理复杂任务(比如自动研究或代码生成),PraisonAI 的低代码设计和多平台集成(如 Telegram)让它上手极快。但作为非官方项目,它的生态成熟度可能不如 LangChain 等主流框架,适合愿意尝鲜的开发者。

AI 与智能体
6.4k

评论