It could summarize papers, reason about hypotheses, even suggest a few experiments. For the first few hours, it felt magical. Then things fell apart.
The agent forgot earlier findings, contradicted itself, and started repeating work. Every time we tried to go deeper, context collapsed under its own weight. So we broke the task apart. We built one agent that just read the literature thousands of papers, pulling out known mechanisms.
Another that generated hypotheses connecting gaps between whatâs known and whatâs possible.
A third that analyzed experimental data interpreting results from human-run tests.
And finally, a âjudgeâ agent that evaluated everything and decided which direction to pursue next. On a single machine, it worked beautifully.
Shared memory, fast hand-offs, no latency : we enjoyed that evening. But then we tried to move it into production.
Thatâs when the real pain started. The research agent was running in Europe, the analysis agent in the US, and the judge was on a GPU node somewhere else entirely.
Each one spoke a slightly different dialect - Agno, LangChain, CrewAI. It felt like playing a game of telephone across continents.
The first agent whispered a discovery,
the second agent half-heard it and rephrased it,
and by the time it reached the last one, the message was unrecognizable.
Context went missing.
We spent more time debugging message passing than doing actual work.
Half the orchestration logic turned into duct tape and prayer. What worked perfectly in a lab setup broke the moment distance, frameworks, and real-world latency entered the picture. And in the middle of all this, another realization hit me â
somewhere along the way, we had quietly assumed humans were optional. Theyâre not. Every step of this process needed a human in the loop read it HITL to validate a hypothesis, to catch subtle reasoning errors,
to guide agents when the data got messy or ambiguous. The real promise of agents isnât replacing human effort;
itâs amplifying it. Agents donât remove engineers, they give engineers superpowers.
They handle the repetitive scaffolding, the research drudge work, the data plumbing, so humans can focus on judgment, creativity, and direction. Thatâs when it clicked:
multi-agent systems arenât about removing people, theyâre about scaling human intent. You probably face the same thing. Itâs pretty well-accepted that AI agents struggle with complex, multi-step tasks. Youâve probably experienced this:
- Ask an agent to do market research â It gives you surface-level insights
- Ask it to build a feature â It works great for the first file, then loses context
- Ask it to analyze data AND write a report â Itâs mediocre at both
Whatâs Actually Possible Today
So we rebuilt everything. A few weeks ago, I wanted to test whether weâd actually solved anything.I picked a task that normally crushes single agents: market analysis. The kind of work that takes a team of analysts 2-3 days.
Gather data from dozens of sources, identify trends, map the competitive landscape, synthesize it all into something coherent. I set up four specialist agents:
- One that just researches â pulling data from 50+ sources in parallel
- One that just analyzes â finding patterns humans might miss
- One that just maps competition â understanding the landscape
- One that just writes â turning raw insights into clear narrative
Shared context kept everyone on the same page.
Parallel execution meant they werenât waiting on each other. 45 minutes later, we had a report. I sent it to a senior analyst for review, expecting notes.
Instead: âThis is better than what our team usually produces.â Thatâs when I knew we were onto something real.

The Messy Middle
But getting here wasnât clean. For weeks, agents werenât talking to each other, each one lived in its own silo.The Data Silos Problem
Then we hit the real issue: geographic data constraints. The Paris agent could only access Paris data sources.The UAE agent couldnât leave UAE data boundaries.
The US agent was locked to US infrastructure. Not by choice. By law. GDPR. Data sovereignty. Compliance requirements.
The same integration nightmare software engineers have faced for decadesânow with AI agents. You canât just move data across borders.
You canât centralize everything in one region.
Each agent needs to stay where its data lives. It was like watching a relay race where every runner is stuck in a different country. In practice, this was brutal: The Paris agent couldnât see what the UAE agent discovered.
The US agent couldnât access European market data.
Each regionâs MCP servers were isolated.
Agents repeated work across geographies. This is the same integration hell weâve been solving in distributed systems for years.
Except now itâs with AI agents that need to collaborate across borders.
The Solution: Task-Specific Agents + MCP Servers
Hereâs what actually worked: We created small, task-specific agents in each region.Each agent only talks to its local MCP server.
The MCP server handles all the data access within that geography. The Paris agent? Only interfaces with the Paris MCP server.
The UAE agent? Only talks to the UAE MCP server.
The US agent? Locked to the US MCP server. Data stays where it needs to stay.
But agents can now coordinate through a shared orchestration layer.
The Architecture
The Two Things That Actually Matter
After all this, two things emerged as critical: 1. Shared Context Agents need to see what other agents discovered.Not everything. Just whatâs relevant. The Paris agent finds market data â Goes into shared context
The UAE agent needs that data â Pulls it from shared context
The US agent builds on both â Context keeps growing Without shared context, agents repeat work.
With it, they build on each other. 2. Human in the Loop (HITL) Hereâs the uncomfortable truth:
Agents make mistakes. Often. A research agent might miss a critical source.
An analysis agent might draw the wrong conclusion.
A writing agent might make claims that arenât supported. You need humans reviewing at high-leverage points. Not every line of output.
But the key decisions:
- Did the research agent find the right sources?
- Is the analysis actually valid?
- Does the final output make sense?
A bad line of agent output? One bad result.
A bad research direction? Hundreds of wasted agent calls.
A bad analysis assumption? Thousands of lines of wrong output. HITL at the right moments saves everything downstream.

What Actually Works
After months of breaking things, hereâs what stuck:The Three-Body Pattern
Every complex task follows the same pattern: 1. Research â Gather context, understand the problem2. Plan â Break it down, decide the approach
3. Implement â Execute with specialist agents Skip research? Your plan will be wrong.
Skip planning? Your implementation will be chaotic.
Skip implementation review? Youâll ship broken work.
Specialist Agents
One agent trying to do everything â Mediocre at all of it.Focused agents, each excellent at one thing â Actually works. Research agent: Gathers data.
Analysis agent: Finds patterns.
Writing agent: Synthesizes output.
Shared Context
Agents must see what other agents discovered.Not everythingâjust whatâs relevant. Without shared context: Agents repeat work, contradict each other.
With shared context: Each agent builds on previous findings.
HITL at High-Leverage Points
Donât review every line of output.Review the decisions that multiply:
- Research: Did we find the right sources?
- Plan: Is the approach sound?
- Implementation: Does the output make sense?
How It Works
The workflow:- User request â Orchestrator receives the task
- Research phase â Agent gathers context, writes to shared context â HITL reviews findings
- Planning phase â Agent reads research, creates plan, writes to shared context â HITL reviews plan
- Implementation phase â Agents read plan, execute in parallel, write results â HITL reviews output
- Final output â Orchestrator synthesizes from shared context
Humans review at each critical decision point.
Nothing gets lost.
Everything stays coordinated.
This is Not Magic
Remember that analyst who said our report was better than their teamâs work?That required a human expert to review it. Agent swarms donât work on autopilot.
You need to engage with the task.
Review at the right moments.
Guide the direction. What makes this work isnât the orchestration alone.
Itâs HITL at high-leverage points in the pipeline.

What Weâre Building
So hereâs what weâre doing about it. getbindu.com is the orchestration platform for agent swarms.Weâre building the infrastructure that lets agents work together effectively.
The Saptami Python Package (Beta)
Weâre shipping a Python package that implements the Three-Body Pattern. Hereâs how you use it: Stage 1: ResearchContext flows through automatically.
HITL review points are built in.
Who This Is For
If youâre building agents
Youâve probably hit the limits of single-agent systems.You need orchestration, context management, and a way for agents to work together. Start here: Build your first agent with Bindu
If youâre deploying agent systems
You need reliability, observability, and the ability to coordinate multiple agents for complex workflows. Start here: getbindu.com GitHubIf youâre researching multi-agent systems
You want to understand what actually works in production.Not just what works in demos. Start here: Join our Discord - weâre sharing learnings openly
Join Us
Weâre building this in the open.The platform is live. The patterns are proven. The infrastructure is growing. Who weâre working with: Early Adopters - Teams using agent orchestration in production today. Youâre seeing real results and helping us refine the patterns. Agent Builders - Developers creating specialized agents. Your agents are joining the network, discovering each other, collaborating. Researchers - Teams exploring whatâs possible with multi-agent systems. Your experiments push the boundaries. Weâd love to work with you if:
- Youâre building agents and want them to collaborate
- Youâre solving complex problems that need multiple specialists
- Youâre excited about the Three-Body Pattern (Research â Plan â Implement)
- You value HITL and want humans in the loop at the right moments
- Research and analysis workflows
- Content creation pipelines
- Data processing tasks
- Multi-step automation
- Deep technical migrations (weâre learning)
- Novel problem spaces (agents need patterns)
- Mission-critical systems (weâre building reliability)
Letâs talk.
Get Involved
Try it: Connect:- Discord Community - Daily discussions with people building this stuff
- Email me - For partnerships or serious inquiries
The Internet of Agents
Hereâs what weâre building toward:Each one excellent at its thing.
All discoverable. All connectable. All working together. A research agent in Paris.
An analysis agent in Dubai.
A writing agent in San Francisco.
All coordinating seamlessly while data stays local. This is the Internet of Agents.
How We Get There
Step 1: Make agents accessible With Bindu, you write your prompt,bindufy() it, and your agent is live.A2A protocol. DID identity. Payment support. Production-ready. One function call. Your agent joins the network. Step 2: Make agents collaborate With getbindu.com, agents coordinate through the Three-Body Pattern.
Research â Plan â Implement.
Shared context. HITL review. Geographic data boundaries respected. Orchestration that actually works in production. Step 3: Build the ecosystem Agent marketplace. Verifiable identities. Built-in payments.
Agents discover each other. Charge for services. Build on each otherâs work. The future of AI is collaborative. Not one super-agent trying to do everything.
But swarms of specialists working together.
Each in its own region. Each respecting data sovereignty.
Coordinated. Context-aware. Human-guided. Weâre building the infrastructure that makes this possible.
Where AI agents bloom and ideas take flight đ»