diff --git a/README.md b/README.md index 3f4281c4..12fdb554 100644 --- a/README.md +++ b/README.md @@ -7,6 +7,7 @@ Memoh

Memoh

Multi-Member, Structured Long-Memory, Containerized AI Agent System.

+

📌 Introduction to Memoh - The Case for an Always-On, Containerized Home Agent

Version License @@ -23,6 +24,8 @@
+ + Memoh is a AI agent system platform. Users can create their own AI bots and chat with them via Telegram, Discord, Lark(Feishu), etc. Every bot has independent container and memory system which allows them to edit files, execute commands and build themselves - Like [OpenClaw](https://openclaw.ai), Memoh provides a more secure, flexible and scalable solution for multi-bot management. ## Quick Start diff --git a/README_CN.md b/README_CN.md index b8df929e..ab117fd6 100644 --- a/README_CN.md +++ b/README_CN.md @@ -7,6 +7,7 @@ Memoh

Memoh

ć€šç”šæˆ·ă€ç»“æž„ćŒ–èź°ćż†ă€ćźč晹挖的 AI Agent çł»ç»Ÿă€‚

+

📌 Introduction to Memoh - The Case for an Always-On, Containerized Home Agent

Version License diff --git a/docs/docs/blogs/2026-02-16.md b/docs/docs/blogs/2026-02-16.md index 64e32f5b..c29e07f8 100644 --- a/docs/docs/blogs/2026-02-16.md +++ b/docs/docs/blogs/2026-02-16.md @@ -6,14 +6,16 @@ author: Team Memoh # Introduction to Memoh - The Case for an Always-On, Containerized Home Agent ## Overview -We enter 2026 with a familiar tension: models get smarter every quarter, but the “agent experience” still breaks on context, latency, privacy, and real-world workflows. Over the past year, We kept circling three questions: -- Where does the capability boundary of agents actually sit? -- What’s the real value of long context? -- What hardware form factor makes “always-on, personal AI” feel natural? + +We enter 2026 with a familiar tension: models get smarter every quarter, but the “agent experience” still breaks on context, latency, privacy, and real-world workflows. Over the past year, we kept circling three questions: +- Where does the capability boundary of agents actually sit? +- What’s the real value of long context? +- What hardware form factor makes “always-on, personal AI” feel natural? Memoh is our attempt to turn those questions into something buildable—not a manifesto, but a system that can survive contact with reality. ## Story Time + Time travels fast. Somewhere between “I’ll remember this” and “wait, why did we decide that?”, a year disappears. That’s the annoying part of building: most progress doesn’t feel like progress while it’s happening. It’s just a stream of small choices, half-finished threads, late-night fixes, and the occasional moment that actually clicks. The kind of moment where you sit back and think: okay—this is real. @@ -30,30 +32,37 @@ Because the thing LLMs can’t give you is not “intelligence.” It’s weight That’s when I realized what I wanted wasn’t “an AI that can talk.” I wanted an AI that can live with you—quietly, continuously, accumulating context without turning your life into content sludge. -Phones were out first instinct—it's personal, powerful, always there. But mobile OS is closed: without OEM privileges you can build an app, not ambient infrastructure. +Phones were our first instinct—it's personal, powerful, always there. But mobile OS is closed: without OEM privileges you can build an app, not ambient infrastructure. -So We looked for the always-on node every home already has: the router (conceptually). Then the economics clash—router-class hardware can’t carry memory, RAG, tools, and multi-user agents. The device evolves: more RAM/storage, a screen, mic/speaker, tiny battery for take out, portable form. +So we looked for the always-on node every home already has: the router (conceptually). Then the economics clash—router-class hardware can’t carry memory, RAG, tools, and multi-user agents. The device evolves: more RAM/storage, a screen, mic/speaker, tiny battery for take out, portable form. Eventually it stops being a router. It becomes a new category: a home agent base layer. ## What + Memoh is a containerized home/studio AI base layer: cloud-grade model capability paired with local-first memory (knowledge base, RAG/search, conversation history) that stays under your control. ## Why + Long-context models raise the ceiling for agents—but they also make “fully local” expensive and “fully cloud” uncomfortable. People don’t want to re-brief AI every day, and they don’t want their durable context trapped in someone else’s feed. Containerization makes Memoh portable, reproducible, and safe to run as always-on infrastructure—so continuity becomes cheap, private, and dependable. ## How + We run Memoh as a containerized stack: isolated services for storage (files/DB/vector index), retrieval, tool/runtime execution, and the control plane. Inference calls cloud APIs when you need frontier capability; durable memory and indexing stay local. The device acts as an always-on node (router-like, not a router) serving multiple users with strict boundaries: sharing is explicit, private context remains private, and everything is deployable/upgradable as versioned containers. ## Features + - **Multi-bot Management**: Create multiple bots; humans and bots, or bots with each other, can chat privately, in groups, or collaborate. -![Multi-bot Management](/blogs/2026-02-16/01-multi-bots.png) + + ![Multi-bot Management](/blogs/2026-02-16/01-multi-bots.png) - **Containerized**: Each bot runs in its own isolated container. Bots can freely execute commands, edit files, and access the network within their containers—like having their own computer. -![Containerized](/blogs/2026-02-16/02-containerized.png) + + ![Containerized](/blogs/2026-02-16/02-containerized.png) - **Memory Engineering**: Every chat is stored in the database, with the last 24 hours of context loaded by default. Each conversation turn is stored as memory and can be retrieved by bots through semantic search. -![Memory Engineering](/blogs/2026-02-16/03-memory-engineering.png) + + ![Memory Engineering](/blogs/2026-02-16/03-memory-engineering.png) - **Various Platforms**: Supports Telegram, Lark (Feishu), and more. - **Simple and Easy to Use**: Configure bots and settings for Provider, Model, Memory, Channel, MCP, and Skills through a graphical interface—no coding required to set up your own AI bot. @@ -61,18 +70,21 @@ We run Memoh as a containerized stack: isolated services for storage (files/DB/v - More... ## Compare to OpenClaw -We Shared core belief: both Memoh and OpenClaw treat the agent as more than a chatbox—we give the LLM a playground: a real environment where it can remember, use tools, and iterate. + +We share a core belief: both Memoh and OpenClaw treat the agent as more than a chatbox—we give the LLM a playground: a real environment where it can remember, use tools, and iterate. Where Memoh differs: + - Lighter and Faster: built as home/studio infrastructure, can be held in the edge device -- Containerized by default: each bot gets an isolated container (files/commands/network/jobs). -- Hybrid split: cloud inference, local-first memory + indexing. -- Multi-user first: explicit sharing and privacy boundaries, support a2a (Agent2Agent). -- Sustainable: have an experienced team and confidence to push forward and build it. +- **Containerized by default**: each bot gets an isolated container (files/commands/network/jobs) +- **Hybrid split**: cloud inference, local-first memory + indexing +- **Multi-user first**: explicit sharing and privacy boundaries, support a2a (Agent2Agent) +- **Sustainable**: have an experienced team and confidence to push forward and build it ## Conclusion + Memoh is built for one thing: always-on continuity—an AI that stays online, and a memory that stays yours. We keep frontier inference in the cloud, keep durable context local, and run everything as a containerized, always-on stack. If you want an agent that feels less like an app and more like home infrastructure, that’s the bet Memoh is making. -Furthermore, we will continue to operate and permanently open source it, permanently open-source Memoh, making it a product with long impact. \ No newline at end of file +Furthermore, we will continue to operate and permanently open-source Memoh, making it a product with long impact. \ No newline at end of file