Proof of Deployment

Automating entire departments with AI agents: what works and what breaks

How to deploy autonomous agents in legal, sales, operations and HR. Lessons from 10 production systems.

Most companies talking about “adopting AI” are thinking about a website chatbot or a developer copilot. That addresses a narrow slice of the problem. The real leap happens when you automate an entire department: not an isolated task, but the complete workflow of a functional area.

At Odisea, we built exactly that. 10 agent systems in production, covering legal, sales, marketing, research, operations and customer service. 90+ defined agent roles, 13,570 lines of Python, real infrastructure processing real data.

This article documents what works, what breaks and what we learned automating departments from scratch.

The first department we fully automated was legal. The context: Ecuadorian legal research for a legal daemon that needs to analyze legislation, regulation and case law, produce syntheses and feed a task backlog with publishable output.

The system has 10 agents with defined roles: corpus engineer, product architect, compliance specialist, market researcher, domain specialist, and five others handling support and quality control functions.

Each task follows a pipeline with 4 quality gates:

  1. Bad pattern detection: 50+ patterns that identify generic, repetitive or substanceless output. If the content triggers any pattern, it goes back for reprocessing.
  2. Content scoring: scale from 0 to 1. Output below 0.4 is automatically rejected.
  3. Retry limit: maximum 3 reprocessing attempts per task. On the third failure, the task is marked as blocked and escalated for human review.
  4. Source verification: citations verified against real legislation databases. Claims without an identifiable source are removed.

The result: 33 of 37 tasks completed without human intervention. The remaining 4 were blocked by funding dependencies (not by system failure). Operational cost: $20/day.

What we learned: legal agents need aggressive quality gates because language models are especially dangerous when they generate text that looks legally precise but contains factual errors. The bad pattern detection gate was the most important component of the system.

Sales department: 7 agents, 92+ prospects

The second case was sales. A pipeline for a DeFi product (Pan.Tech) with 92+ prospects in a Notion CRM, managed by 7 specialized agents.

The roles include: market researcher, lead enricher, meeting tracker, competitor analyst, proposal generator, pipeline manager and follow-up coordinator.

The flow works like this: new prospects enter Notion via form or manual import. The lead enricher pulls supplementary information (company size, tech stack, funding round, decision-makers). The market researcher cross-references sector data. The competitor analyst maps who else is selling to that prospect. The proposal generator assembles a customized proposal based on the profile. The coordinator schedules follow-ups and tracks responses.

What works: automatic lead enrichment and pipeline tracking deliver the most value. Without them, the team would spend 3-4 hours per week researching each prospect manually. With agents, research happens in minutes and data appears pre-formatted in the CRM.

What breaks: proposal generation needs human review. Agents produce proposals that are structurally correct but miss nuances of commercial relationships. An agent doesn’t know that the CEO of that company was the founder’s college classmate, or that there was an informal meeting at last week’s event. Proposals always go through review before sending.

Operations department: multi-team orchestration

The most complex operation is the orchestration of agent teams. At Odisea, 6 teams with 23+ roles coordinate work in parallel, with managed dependencies and autonomous sprints.

The central mechanism is simple: each task has three levels of authority.

  • T1 (autonomous): research, analysis, memory updates. The agent executes and logs.
  • T2 (notify): outreach, applications, proposals. The agent executes and sends notification.
  • T3 (wait): contracts, terms, launches, hiring. The agent prepares and waits for approval.

Without this hierarchy, automating entire departments is unworkable. Agents with unrestricted authority will, eventually, send an email they shouldn’t have, post content that wasn’t reviewed or agree to terms nobody approved.

The key is designing the boundaries before deployment. Each department has a decision map with risk classification. Low-risk decisions (researching information, formatting data, updating CRM) are T1. Medium-risk decisions (emailing a prospect, publishing a content draft) are T2. High-risk decisions (signing a contract, changing prices, firing someone) are T3.

HR and compliance: where caution is mandatory

HR and compliance departments are the most sensitive for automation. Personal data, labor regulation that varies by country (Brazil’s CLT differs from Mexico’s LFT and Argentina’s legislation), and severe consequences for errors.

Our approach to these departments is deliberately more conservative:

Triage agents, not decision agents. In compliance, an agent can scan documentation, identify gaps, generate checklists and prepare reports. The conformity decision is human. In HR, an agent can process applications, schedule interviews and generate candidate summaries. The hiring decision is human.

Continuous auditing. All output from HR and compliance agents is logged with full traceability: which agent generated it, what input it received, which model was used, when it was generated. This is a requirement for compliance with LGPD and other regional regulations.

Narrow scope by design. Instead of automating “the HR department,” we automate specific tasks: resume screening, interview scheduling, onboarding checklist generation, documentation tracking. Each task has explicit limits and defined escalation points.

The pattern that works

After deploying agents across 6 different departments, the pattern that emerges is consistent:

  1. Start with the backlog, not the org chart. Don’t automate “the legal department.” Automate “the 37 research tasks that have been sitting idle for 3 months because nobody has time.” The actual backlog dictates priorities.

  2. Quality gates before scale. A bad agent scaled to 100 tasks produces 100 bad outputs. Build the gates first, run 5-10 tasks with supervision, calibrate the thresholds, then open the volume.

  3. Integration with existing tools. Agents that live in a parallel system get ignored. Agents that post in the team’s Slack, update the Notion everyone uses and send email from the company domain get adopted.

  4. Explicit authority hierarchy. Every agent knows what it can do on its own, what it does and reports, and what it prepares and waits on. No ambiguity.

  5. Cost-per-task metrics. We know how much each task costs in tokens, processing time and API calls. This lets us compare against the cost of a human doing the same work and demonstrate concrete ROI.

Automating an entire department is not an 18-month project with a consultancy billing by the hour. With the right engineering, the first functional agent enters production in 2 weeks. In 60 days, a department has 5-10 agents operating on real tasks. In 90 days, ROI is measurable.

The question is not whether it’s worth it. It’s which tasks you automate first.


Synaptic transforms companies into AI-native organizations. We start where the demo ends. synaptic.so