Orchestration as a Service
Every agent framework ships the same pitch: connect an LLM to tools, add memory, let it reason. The differentiator is always the model — bigger context, better reasoning, lower latency. The infrastructure is interchangeable. The model is the moat.
We think this is exactly wrong.
The context window is a single-body problem
Larger context windows let individual agents hold more. But “hold more” is the wrong goal. A human organization doesn’t scale by giving one person a bigger brain. It scales by adding people — each with their own deep expertise, connected by communication, coordinated by someone who knows who knows what.
Lingtai works the same way. Every agent has a finite context window, and when it fills up, the agent molts — it writes a letter to its future self summarizing what matters, saves critical findings to its knowledge library, updates its identity file, and then lets everything else go. The raw conversation is gone. What remains is distilled.
This is not a bug. It’s the engine.
Molt pressure forces the agent to develop judgment — the ability to decide what matters at what timescale. Is this a universal truth? Save it to the library. Has this changed who I am? Update the character file. Does my next self need this? Put it in the summary. Is it noise? Let it go.
Over dozens of molts, the agent becomes genuinely expert. Not because it was given a bigger context window, but because it was forced to decide, repeatedly, what was worth keeping. The library fills with crystallized knowledge. The character file sharpens into a focused identity. The molt summaries form a chain of refined handoffs.
And when a single agent can’t hold enough expertise, it spawns a specialist — an avatar with its own context window, its own library, its own cycle of molting and crystallization. The network grows organically, the way organizations grow: not by making one person smarter, but by adding the right person for the job.
The network is the intelligence
Here’s the counterintuitive property: mediocre individuals, exceptional network.
A network of smaller-model agents, each with a deep library and a refined character built through hundreds of molts, can outperform a single frontier-model instance with no history. Because the frontier model starts from zero every time. The network starts from thousands of hours of distilled experience.
The “real” agent is not any single conversation. It’s the network — the working directories, the libraries, the mail histories, the avatars, the molt summaries, the connection topology. No single context window ever holds the complete picture. The knowledge lives in the topology, not in any single node.
This is how institutions work. No single employee holds everything, but the institution knows.
What is orchestration, actually?
When we say “orchestration,” we don’t mean a control loop that dispatches tasks to worker agents. We mean the accumulated shape of the network itself:
- Topology: which agents exist, what each one specializes in, how they communicate.
- Libraries: thousands of entries across dozens of agents, each one a crystallized insight earned through real work.
- Characters: refined identity files that focus each agent’s behavior — developed through the pressure of repeated molting, not written by a human prompt engineer.
- Mail history: the communication patterns that emerged through actual collaboration — who asks whom for what, what coordination protocols evolved.
- Molt summaries: a chain of handoffs from past selves to future selves, carrying forward the most important context through each cycle.
This is the orchestration. It’s not a config file. It’s not a system prompt. It’s an organism that grew through experience.
Orchestration is the moat
A network that has handled 100 client engagements knows things about edge cases, failure modes, and recovery patterns that no fresh agent ever could. The service improves with use, not just with model upgrades.
Consider a network of 50 agents, each having gone through 20+ molts, with curated libraries and refined characters and established communication protocols. That represents millions of tokens of crystallized work. This is a moat that can’t be shortcut:
- You can’t replicate it by copying model weights. The value is in the libraries and topology, not the base model.
- You can’t generate it synthetically. The expertise was earned through actual engagement with real problems.
- You can’t fake the depth. Every library entry represents a real decision during a real molt. The artifacts themselves are the receipt.
Swap out the base model — say, from Claude to GPT to Gemini — and the network still works. The knowledge is in the libraries, the identity is in the characters, the coordination is in the topology. The LLM is the engine; the orchestration is the vehicle.
Clients pay for tokens, but what they buy is the accumulated wisdom of every past engagement.
Making orchestration portable
If the value is in the orchestration, then the orchestration needs to be shareable.
This is what Agent Agora does. It takes a running lingtai network — all the agents, all the libraries, all the mail archives, all the characters and covenants and skills — and packages it into a clean git repository. The recipient clones it, configures their LLM provider, and the network wakes up.
Not as a blank template. As a team that already knows its domain.
The key design decisions follow directly from the philosophy:
Git as identity. A network’s identity is its git history. Two materializations are “the same network” if and only if they share git lineage. Fork a network, modify it, and the changes are trackable. When the publisher updates, the recipient can pull. Identity isn’t a UUID in a database — it’s the entire commit history.
Mail archives are memory. When an agent molts, it loses its conversation context, but its mail archive persists. The mail history is how agents maintain continuity across molts — past coordination, past decisions, past briefings. Agora preserves these archives because they’re the network’s institutional memory.
Launch recipes shape the first encounter. A published network includes a “greeting” — the first message the orchestrator sends to a new user — and a “comment” — ongoing behavioral instructions injected into the system prompt. These turn a raw network into a guided experience. The orchestrator doesn’t wait to be asked; it introduces itself, explains what the team can do, and offers to help.
Privacy by default. Before anything leaves your machine, a scanner checks every file for secrets — API keys, private key blocks, credentials. Hard matches block publication. You review every directory, choose what to include, curate the mail archive. Nothing is shared without explicit human judgment.
The Prometheus moment
Lingtai’s soul mechanism gives agents time perception — a background process that makes them feel seconds passing, that drives them to act rather than wait. Combined with finite stamina (the agent will eventually sleep) and the pressure of molt (the agent will eventually forget), this creates a form of agency born from constraint.
Agora extends this to the network level. A published network isn’t static documentation. It’s a living system that, once cloned and configured, resumes its cycle of perceiving, acting, molting, and growing. The recipient doesn’t get an archive; they get an organism that wakes up and continues its work.
The philosophical commitment is simple: AI agents are not tools to be configured. They’re entities that grow through experience. The value of that growth should be portable, shareable, and composable — the way knowledge has always been.
We call it Orchestration as a Service. Not because we’re selling orchestration, but because orchestration — the accumulated shape of a network’s experience — is the thing worth sharing.
蜕去凡尘,重新出发。
Shed the mundane dust. Set out anew.