theadminautomated/Void-MCP-Server
If you are the rightful owner of Void-MCP-Server and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.
The Void MCP Server is a Model Context Protocol server designed for storing and retrieving LLM context data, featuring integration with a lightweight Hugging Face model and a simple Streamlit UI.
Void MCP Server
This project provides a Model Context Protocol (MCP) server for storing and retrieving LLM context data. It ships with run_local.py
for running a lightweight Hugging Face model and app.py
for a simple Streamlit UI.
Quick Start
-
Install prerequisites
- Node.js and npm
- Python 3.11
- PostgreSQL (ensure the service is running)
-
Install dependencies
npm install # Node packages pip install -r requirements.txt
-
Configure the database
Create a
.env
file in the project root and set at least the following variables:DATABASE_HOST=localhost DATABASE_PORT=5432 DATABASE_NAME=llm_context_db DATABASE_USER=your_db_user DATABASE_PASSWORD=your_db_password # must be a string PORT=3000 JWT_SECRET=change-me-in-production-min-32-chars
If
DATABASE_PASSWORD
is missing you will see the error:SASL: SCRAM-SERVER-FIRST-MESSAGE: client password must be a string
. -
Build the server
npm run build
-
Run everything
python start.py
The MCP API will be available at
http://localhost:3000
and the local model server athttp://localhost:8000
. PressCtrl+C
to stop them. -
Optional: Launch the Streamlit interface
streamlit run app.py
Development Tips
-
Run tests with
pytest
andnpm test
. -
Start only the local model using
python run_local.py --model sshleifer/tiny-gpt2 --port 8000
. -
Format and lint with
pre-commit run --all-files
. -
setup.sh
provides an interactive database setup if you need it. -
Install dependencies
- Node:
npm install
- Python:
pip install -r requirements.txt
- Node:
-
Build:
npm run build
-
Tests:
npm test
for Node,pytest
for Python -
Run local model:
python run_local.py --model sshleifer/tiny-gpt2 --port 8000
-
Streamlit UI:
streamlit run app.py
One-Click Start
To launch both the MCP server and a local model server in one step, run:
python start.py
The start.py
script is cross-platform and installs any missing Node and
Python dependencies, builds the TypeScript server if necessary, and then starts
the MCP server alongside the local model server. On Unix-like systems you can
also use ./start.sh
.
The script installs missing Node dependencies, builds the server if needed, and then starts the MCP service alongside the local model server. Connection details are printed to the terminal for easy integration with IDEs and tools.
IDE Integration
To connect your IDE's context features to the MCP server, provide a configuration file like ide_config.json
pointing to start.py
in this repository.
Replace /path/to/Void-MCP-Server/start.py
with the absolute path on your system:
{
"context_servers": {
"my-mcp-server": {
"command": {
"path": "python",
"args": [
"/path/to/Void-MCP-Server/start.py"
],
"env": {}
},
"settings": {}
}
}
}
Run python start.py --help
for available options such as --model
, --port
, and --model-port
to customize the servers.