mirror of
https://github.com/punkpeye/awesome-mcp-servers.git
synced 2026-03-12 04:05:17 +00:00
Add Glama link for remembra
This commit is contained in:
@@ -1160,7 +1160,7 @@ Persistent memory storage using knowledge graph structures. Enables AI models to
|
||||
- [pinecone-io/assistant-mcp](https://github.com/pinecone-io/assistant-mcp) 🎖️ 🦀 ☁️ - Connects to your Pinecone Assistant and gives the agent context from its knowledge engine.
|
||||
- [pomazanbohdan/memory-mcp-1file](https://github.com/pomazanbohdan/memory-mcp-1file) 🦀 🏠 🍎 🪟 🐧 - A self-contained Memory server with single-binary architecture (embedded DB & models, no dependencies). Provides persistent semantic and graph-based memory for AI agents.
|
||||
- [ragieai/mcp-server](https://github.com/ragieai/ragie-mcp-server) 📇 ☁️ - Retrieve context from your [Ragie](https://www.ragie.ai) (RAG) knowledge base connected to integrations like Google Drive, Notion, JIRA and more.
|
||||
- [remembra-ai/remembra](https://github.com/remembra-ai/remembra) 🐍 📇 🏠 ☁️ 🍎 🪟 🐧 - Persistent memory layer for AI agents with entity resolution, PII detection, AES-256-GCM encryption at rest, and hybrid search. 100% on LoCoMo benchmark. Self-hosted.
|
||||
- [remembra-ai/remembra](https://github.com/remembra-ai/remembra) ([glama](https://glama.ai/mcp/servers/remembra-ai/remembra)) 🐍 📇 🏠 ☁️ 🍎 🪟 🐧 - Persistent memory layer for AI agents with entity resolution, PII detection, AES-256-GCM encryption at rest, and hybrid search. 100% on LoCoMo benchmark. Self-hosted.
|
||||
- [redleaves/context-keeper](https://github.com/redleaves/context-keeper) 🏎️ 🏠 ☁️ 🍎 🪟 🐧 - LLM-driven context and memory management with wide-recall + precise-reranking RAG architecture. Features multi-dimensional retrieval (vector/timeline/knowledge graph), short/long-term memory, and complete MCP support (HTTP/WebSocket/SSE).
|
||||
- [roomi-fields/notebooklm-mcp](https://github.com/roomi-fields/notebooklm-mcp) ([glama](https://glama.ai/mcp/servers/@roomi-fields/notebooklm-mcp)) 📇 🏠 🍎 🪟 🐧 - Full automation of Google NotebookLM — Q&A with citations, audio podcasts, video, content generation, source management, and notebook library. MCP + HTTP REST API.
|
||||
- [rushikeshmore/CodeCortex](https://github.com/rushikeshmore/CodeCortex) [glama](https://glama.ai/mcp/servers/@rushikeshmore/codecortex) 📇 🏠 🍎 🪟 🐧 - Persistent codebase knowledge layer for AI coding agents. Pre-digests codebases into structured knowledge (symbols, dependency graphs, co-change patterns, architectural decisions) via tree-sitter native parsing (28 languages) and serves via MCP. 14 tools, ~85% token reduction. Works with Claude Code, Cursor, Codex, and any MCP client.
|
||||
|
||||
Reference in New Issue
Block a user