MARM-Systems

Lyellr88/MARM-Systems

3.6

If you are the rightful owner of MARM-Systems and would like to certify it and/or have it hosted online, please leave a comment on the right or send an email to henry@mcphub.com.

MARM (Memory Accurate Response Mode) is a comprehensive AI memory ecosystem designed to solve the problem of context loss in large language models.

Tools
5
Resources
0
Prompts
0
MARM - The AI That Remembers Your Conversations

MARM: The AI That Remembers Your Conversations

Memory Accurate Response Mode v2.1.0 - The intelligent memory system for AI agents. Stop losing context. Stop hallucinations. Start controlling your LLM conversations.

Stars Forks Version License VersionPython FastAPI

Official MARM

Note: This is the official MARM repository. All official versions and releases are managed here.

Forks may experiment, but official updates will always come from this repo.


πŸš€ Quick Start for MCP

Docker (Fastest - 30 seconds):

docker pull lyellr88/marmcp-beta:latest
docker run -d --name marm-mcp-server -p 8001:8001 lyellr88/marmcp-beta:latest
claude mcp add marm-memory http://localhost:8001/mcp

Quick Local Install:

git clone https://github.com/MARM-Systems/MARM.git
cd MARM/marm-mcp-server/MARMcp-beta
# Unix/Mac: ./install.sh | Windows: python setup.py
claude mcp add marm-memory http://localhost:8001/mcp

Key Information:

  • Server Endpoint: http://localhost:8001/mcp
  • API Documentation: http://localhost:8001/docs
  • Supported Clients: Claude Code, Qwen CLI, Gemini CLI, and any MCP-compatible LLM client or LLM platform

All Installation Options:

  • Docker (Fastest): One command, works everywhere
  • Automated Setup: One command with dependency validation
  • Manual Installation: Step-by-step with virtual environment
  • Quick Test: Zero-configuration trial run

Choose your installation method:

Installation TypeGuideBest For
DockerINSTALL-DOCKER.mdCross-platform, production deployment
WindowsINSTALL-WINDOWS.mdNative Windows development
LinuxINSTALL-LINUX.mdNative Linux development
PlatformsINSTALL-PLATFORM.mdApp & API integration

πŸš€ Quick Start for Chatbot

The MARM Chatbot is a web-based interface for interacting with the MARM protocol. To install it, you'll need Node.js and a Replicate API key.

1. Clone the repository

git clone https://github.com/MARM-Systems/MARM.git

2. Install dependencies

cd MARM-Systems/webchat
npm install

3. Add your Replicate API key to a .env file

echo "REPLICATE_API_TOKEN=your_replicate_api_token_here" > .env

4. Start the server

npm start

Fastest MARM Chatbot


🎯 Why MARM?

MARM (Memory Accurate Response Mode) is a comprehensive AI memory ecosystem I designed to solve the problem of context loss in large language models. What started as a simple protocol has evolved into a suite of tools that provide a persistent, intelligent, and cross-platform memory for any AI agent.

The MARM ecosystem consists of three main components:

  • The MARM Protocol: A set of rules and commands for structured, reliable AI interaction.
  • The MARM Universal MCP Server: A production-ready memory intelligence platform that provides a powerful, stateful backend for any MCP-compatible AI client.
  • The MARM Chatbot: A web-based interface for interacting with the MARM protocol directly.

Whether you're a developer looking to build the next generation of AI agents, a researcher studying AI behavior, or simply a power user who wants to have more productive conversations with your AI, the MARM ecosystem provides the tools you need to unlock the full potential of large language models.

MARM - The AI That Remembers Your Conversations

*Appears in Google AI Overview for AI memory protocol queries (as of Aug 2025)*

The newest addition tho the ecosystem is MARM MCP it represents an emerging category of MCP server that integrates a complete protocol layer with intelligent memory systems. Built on FastAPI and SQLite, it combines the MARM protocol with semantic search, session management, and smart retrieval to bridge tool access with structured reasoning. This creates a more consistent, user-controlled LLM experience that goes beyond simple tool exposure.

CategoryFeatureDescription
🧠 MemorySemantic SearchFind memories by meaning using AI embeddings, not keyword matching
Auto-ClassificationContent intelligently categorized (code, project, book, general)
Cross-Session MemoryMemories survive across different AI agent conversations
Smart RecallVector similarity search with context-aware intelligent fallbacks
🀝 Multi-AIUnified Memory LayerAccessible by any connected LLM (Claude, Qwen, Gemini, etc.)
Cross-Platform IntelligenceDifferent AI agents learn from each other's interactions
User-Controlled MemoryGranular control over memory sharing and "Bring Your Own History"
πŸ—οΈ Architecture19 Complete MCP ToolsFull Model Context Protocol implementation
Database OptimizationSQLite with WAL mode and connection pooling
Rate LimitingIP-based protection for sustainable free service
MCP ComplianceResponse size management for optimal performance
Docker ReadyContainerized deployment with health monitoring
⚑ AdvancedUsage AnalyticsPrivacy-conscious insights for platform optimization
Event-Driven SystemSelf-managing architecture with comprehensive error isolation
Structured LoggingDevelopment and debugging support with structlog
Health MonitoringReal-time system status and performance tracking

Why I Built MARM

MARM started with my own frustrations: AI losing context, repeating itself, and drifting off track. But I didn’t stop there. I asked a simple question in a few AI subreddits:
β€œWhat’s the one thing you wish your LLM could do better?”

The replies echoed the same pain points:

  • Keep memory accurate
  • Give users more control
  • Be transparent, not a black box

That feedback confirmed the gap I already saw. I took those shared frustrations, found the middle ground, and built MARM. Early contributors validated the idea and shaped features, but the core system grew out of both personal trial and community insight.

MARM is the result of combining individual persistence with collective needs, a protocol designed to solve what we all kept running into.

Discord

Join Discord for upcoming features and builds, plus a safe space to share your work and get constructive feedback.

MARM Discord


Before MARM vs After MARM

Without MARM:

  • "Wait, what were we discussing about the database schema?"
  • AI repeats previous suggestions you already rejected
  • Loses track of project requirements mid-conversation
  • Starts from scratch every time you return

With MARM:

  • AI references your logged project notes and decisions
  • Maintains context across multiple sessions
  • Builds on previous discussions instead of starting over
  • Remembers what works and what doesn't for your project

Why Use MARM?

Modern LLMs often lose context or fabricate information. MARM introduces a session memory kernel, structured logs, and a user-controlled knowledge library. Anchoring the AI to your logic and data. It’s more than a chatbot wrapper. It’s a methodology for accountable AI.

Command Overview

CategoryCommandFunction
Session/start marmActivate protocol
/refresh marmReaffirm/reset context
Core/logStart structured session logging
/notebookStore key data
/summary:Summarize and reseed sessions
Advanced/deep diveRequest context-aware response
/show reasoningReveal logic trail of last answer

Need a walkthrough or troubleshooting help? The MARM-HANDBOOK.md covers all aspects of using MARM.


πŸš€ Try MARM Chatbot Now - No Setup Required

Experience all features instantly in your browser

Need detailed steps, troubleshooting, or multi-provider setup?

See CHATBOT-SETUP.md for complete installation guide with Node.js setup and troubleshooting.

πŸ› οΈ MARM MCP Server Guide

Now that you understand the ecosystem, here's info and how to actually use the MCP server with your AI agents


πŸ› οΈ Complete MCP Tool Suite (19 Tools)

CategoryToolDescription
🧠 Memory Intelligencemarm_smart_recallAI-powered semantic similarity search across all memories. Supports global search with search_all=True flag
marm_contextual_logIntelligent auto-classifying memory storage using vector embeddings
πŸš€ Session Managementmarm_startActivate MARM intelligent memory and response accuracy layers
marm_refreshRefresh AI agent session state and reaffirm protocol adherence
πŸ“š Logging Systemmarm_log_sessionCreate or switch to named session container
marm_log_entryAdd structured log entry with auto-date formatting
marm_log_showDisplay all entries and sessions (filterable)
marm_log_deleteDelete specified session or individual entries
πŸ”„ Reasoning & Workflowmarm_summaryGenerate context-aware summaries with intelligent truncation for LLM conversations
marm_context_bridgeSmart context bridging for seamless AI agent workflow transitions
πŸ“” Notebook Managementmarm_notebook_addAdd new notebook entry with semantic embeddings
marm_notebook_useActivate entries as instructions (comma-separated)
marm_notebook_showDisplay all saved keys and summaries
marm_notebook_deleteDelete specific notebook entry
marm_notebook_clearClear the active instruction list
marm_notebook_statusShow current active instruction list
βš™οΈ System Utilitiesmarm_current_contextGet current date/time for accurate log entry timestamps
marm_system_infoComprehensive system information, health status, and loaded docs
marm_reload_docsReload documentation into memory system

πŸ—οΈ Architecture Overview

Core Technology Stack

FastAPI (0.115.4) + FastAPI-MCP (0.4.0) - v2.1.0
β”œβ”€β”€ SQLite with WAL Mode + Custom Connection Pooling  
β”œβ”€β”€ Sentence Transformers (all-MiniLM-L6-v2) + Semantic Search
β”œβ”€β”€ Structured Logging (structlog) + Memory Monitoring (psutil)
β”œβ”€β”€ IP-Based Rate Limiting + Usage Analytics
β”œβ”€β”€ MCP Response Size Compliance (1MB limit)
β”œβ”€β”€ Event-Driven Automation System
β”œβ”€β”€ Docker Containerized Deployment + Health Monitoring
└── Advanced Memory Intelligence + Auto-Classification

Database Schema (5 Tables)

memories - Core Memory Storage
CREATE TABLE memories (
    id TEXT PRIMARY KEY,
    session_name TEXT NOT NULL,
    content TEXT NOT NULL,
    embedding BLOB,              -- AI vector embeddings for semantic search
    timestamp TEXT NOT NULL,
    context_type TEXT DEFAULT 'general',  -- Auto-classified content type
    metadata TEXT DEFAULT '{}',
    created_at TEXT DEFAULT CURRENT_TIMESTAMP
);
sessions - Session Management
CREATE TABLE sessions (
    session_name TEXT PRIMARY KEY,
    marm_active BOOLEAN DEFAULT FALSE,
    created_at TEXT DEFAULT CURRENT_TIMESTAMP,
    last_accessed TEXT DEFAULT CURRENT_TIMESTAMP,
    metadata TEXT DEFAULT '{}'
);
Plus: log_entries, notebook_entries, user_settings

πŸ“ˆ Performance & Scalability

Production Optimizations

  • Custom SQLite Connection Pool: Thread-safe with configurable limits (default: 5)
  • WAL Mode: Write-Ahead Logging for concurrent access performance
  • Lazy Loading: Semantic models loaded only when needed (resource efficient)
  • Intelligent Caching: Memory usage optimization with cleanup cycles
  • Response Size Management: MCP 1MB compliance with smart truncation

Rate Limiting Tiers

  • Default: 60 requests/minute, 5min cooldown
  • Memory Heavy: 20 requests/minute, 10min cooldown (semantic search)
  • Search Operations: 30 requests/minute, 5min cooldown

πŸ“š Documentation for MCP

Guide TypeDocumentDescription
Docker SetupINSTALL-DOCKER.mdCross-platform, production deployment
Windows SetupINSTALL-WINDOWS.mdNative Windows development
Linux SetupINSTALL-LINUX.mdNative Linux development
Platform IntegrationINSTALL-PLATFORM.mdApp & API integration
MCP HandbookMCP-HANDBOOK.mdComplete usage guide with all 19 MCP tools, cross-app memory strategies, pro tips, and FAQ

πŸ†š Competitive Advantage

vs. Basic MCP Implementations

FeatureMARM v2.1.0Basic MCP Servers
Memory IntelligenceAI-powered semantic search with auto-classificationBasic key-value storage
Tool Coverage19 complete MCP protocol tools3-5 basic wrappers
ScalabilityDatabase optimization + connection poolingSingle connection
MCP Compliance1MB response size managementNo size controls
DeploymentDocker containerization + health monitoringLocal development only
AnalyticsUsage tracking + business intelligenceNo tracking
Codebase Maturity2,500+ lines professional code200-800 lines

🀝 Contributing

Aren't you sick of explaining every project you're working on to every LLM you work with?

MARM is building the solution to this. Support now to join a growing ecosystem - this is just Phase 1 of a 3-part roadmap and our next build will complement MARM like peanut butter and jelly.

Join the repo that's working to give YOU control over what is remembered and how it's remembered.

Why Contribute Now?

  • Ground floor opportunity - Be part of the MCP memory revolution from the beginning
  • Real impact - Your contributions directly solve problems you face daily with AI agents
  • Growing ecosystem - Help build the infrastructure that will power tomorrow's AI workflows
  • Phase 1 complete - Proven foundation ready for the next breakthrough features

Development Priorities

  1. Load Testing: Validate deployment performance under real AI workloads
  2. Documentation: Expand API documentation and LLM integration guides
  3. Performance: AI model caching and memory optimization
  4. Features: Additional MCP protocol tools and multi-tenant capabilities

Join the MARM Community

Help build the future of AI memory - no coding required!

Connect: MARM Discord | GitHub Discussions

Easy Ways to Get Involved

  • Try the MCP server or Chatbot and share your experience
  • Star the repo if MARM solves a problem for you
  • Share on social - help others discover memory-enhanced AI
  • Open issues with bugs, feature requests, or use cases
  • Join discussions about AI reliability and memory

For Developers

  • Build integrations - MCP tools, browser extensions, API wrappers
  • Enhance the memory system - improve semantic search and storage
  • Expand platform support - new deployment targets and integrations
  • Submit Pull Requests - Every PR helps MARM grow. Big or small, I review each with respect and openness to see how it can improve the project

⭐ Star the Project

If MARM helps with your AI memory needs, please star the repository to support development!


Star History Chart


License & Usage Notice

This project is licensed under the MIT License. Forks and derivative works are permitted.

However, use of the MARM name and version numbering is reserved for releases from the official MARM repository.

Derivatives should clearly indicate they are unofficial or experimental.


πŸ“ Project Documentation

Usage Guides

  • MARM-HANDBOOK.md - Original MARM protocol handbook for chatbot usage
  • MCP-HANDBOOK.md - Complete MCP server usage guide with commands, workflows, and examples
  • PROTOCOL.md - Quick start commands and protocol reference
  • FAQ.md - Answers to common questions about using MARM

MCP Server Installation

Chatbot Installation

Project Information


Built with ❀️ by MARM Systems - Universal MCP memory intelligence