I've been building AI agent pipelines for the past year and curious how others handle it. Specifically:
- Do you use a framework (LangChain, CrewAI) or roll your own
- How do you handle agent-to-agent data passing?
- What does your observability look like for agent runs?
- Are you running agents on cron/webhooks or manual-only?
Interested in hearing what's working and what's painful.
The naive approach is stateless. Each reply gets processed independently. This breaks down fast when a prospect says "as I mentioned before" and the agent has no memory of what they mentioned before.
What has worked better: treating the entire conversation thread as the context window, not just the latest message. Every reply, every prior message, the research done on the prospect at the start, all of it gets passed through. The agent always knows where it is in the conversation and what has already been said.
The second problem is confidence calibration. Multi-agent systems in production need to know when to act autonomously and when to surface something for human review. In sales specifically, the cost of an agent saying something wrong to a real prospect is high. We err toward flagging ambiguous situations rather than guessing.
The pattern that has held up: agents own clearly bounded tasks end to end (research, draft, send, parse reply), with a thin orchestration layer that routes based on reply classification. Classification is the hardest part to get right and the most important to get right.
reply