Jongbin-kr/MyArxivDB_MCP
If you are the rightful owner of MyArxivDB_MCP and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.
This repository provides a MCP server to organize personal research papers using an MCP server and semantic search.
π§ Organizing Research Papers with MCP Server
This repository provides a MCP server to organize personal research papers using an MCP (Modular Command Platform) server and semantic search. The platform enables efficient paper retrieval, automatic project assignment, and assistance in writing literature review sections using LLMs.
This server is mainly designed for Claude Desktop but may also work well with other MCP clients.
This project was carried out as part of the term project for the BKMS1 course(@SNU GSDS).
β‘οΈ Quickstart
The following quickstart guide is based on an Apple Silicon MacBook.
- Install uv
curl -LsSf https://astral.sh/uv/install.sh | sh
- clone the repository
git clone https://github.com/Jongbin-kr/MyArxivDB_MCP.git
- Intsall dependencies & activate the virtual environment
uv sync
source /.venv/bin/activate
- set up environment variables at
.env
file.
# .env
PINECONE_API_KEY = "YOUR PINECONE_API_KEY"
DB_NAME = "YOUR_DB_NAME"
DB_USER = "YOUR_DB_USERNAME"
DB_PASSWORD = "YOUR_DB_PASSWORD"
DB_HOST = "localhost"
DB_PORT = 4444
-
Intsall MCP server at Claude Desktop
mcp install server.py
-
Done! The Claude desktop app will automatically detect the MCP server and you can start using i!
π Motivation
Researchers frequently accumulate large numbers of papers but lack tools to systematically organize them by topic or project. BKMS aims to:
- Automatically assign new papers to relevant projects using embeddings
- Allow semantic search for project-specific literature
- Assist in drafting sections like βRelated Workβ using LLMs
π οΈ Main functions
Our MCP server supports the following core capabilities:
- Crawling metadata and PDFs from arXiv using ID or URL using ArXiv API
- Embedding abstracts using Pinecone API(
llama-text-embed-v2
) - Storing papers and projects in a PostgreSQL + pgvector DB
- Generating "Related Work" sections via LLM prompts
π₯ Workflow & Demo video
You can see our PPT and demo video in assets folder.
Brief overview of our project workflow and DB schema is as follows.
π¨βπ©βπ§βπ¦ Team
- λ°μ°μ§
- μμ’ λΉ
- μ μμ€