The AI-First BPO: A Strategic Blueprint for Building the Next-Generation Outsourcing Company
Trilled — Mexico City April 2026
Executive Summary
The $400B+ global BPO industry is undergoing its most significant transformation since offshoring went mainstream in the 2000s. The shift from labor arbitrage to AI-augmented service delivery is not a future trend — it is happening now. TaskUs is actively cannibalizing its own BPO revenue to build an AI moat. Concentrix is pivoting its entire business model. Yet a massive gap remains: small and medium businesses (SMBs) are not ready to adopt AI directly.
Research shows 43% of SMBs cite lack of expertise as their primary barrier to AI adoption, 38% struggle with integration, and 82% of the smallest businesses believe AI simply does not apply to them. Meanwhile, 57% of SMBs increased AI investment in 2025, up from 36% in 2023 — demand is accelerating, but capability is not.
This creates a structural opportunity for Trilled: position as a traditional BPO/nearshore partner, win clients with familiar per-seat or per-hour pricing, learn the client’s business deeply, then progressively introduce AI agents that augment and eventually replace manual processes. The result is a cash-flowing AI business funded by BPO revenue — with margins that improve over time rather than erode.
This white paper lays out the market opportunity, the hybrid human+AI operating model, the technical architecture, and a phased implementation roadmap to execute this strategy from Mexico City.
Key numbers: - Global BPO market: $406B (2025), projected $733B by 2033 (10.1% CAGR) - AI agent cost per resolution: $0.50-$1.25 vs. human agent: $4-$8 - TaskUs AI services: growing 60.8% YoY, targeting 25% of revenue by 2026 - SMB AI adoption barrier: 43% lack expertise, creating outsourcing demand - Mexico nearshore savings: 40-60% vs. U.S. labor costs - Gartner warning: GenAI cost per resolution will exceed $3 by 2030 — hybrid model wins
Table of Contents
- Market Landscape
- The Hybrid Human+AI Model
- Revenue Model Evolution
- Client Acquisition Strategy
- Architecture for Customer Support AI Agents
- Knowledge Management and Ingestion
- Conversation Design and Quality
- Integration Patterns
- Data Access & Tooling Strategy
- Workforce Career Progression Model
- Training and Continuous Improvement
- Safety, Compliance, and Guardrails
- AI for Recruiting Operations
- Mexico-Specific Recruiting
- LLM Selection Guide
- Agent Framework Comparison
- Evaluation and Monitoring
- Cost Modeling
- Implementation Roadmap
- Accelerated Roadmap (AI-Native Engineering)
- Risks and Mitigations
1. Market Landscape
Industry Size and Growth
The global BPO market was valued at $406 billion in 2025 and is projected to reach $733 billion by 2033, growing at a CAGR of 10.1%. The market is expanding from $436B in 2026 to $623B by 2031 at 7.4% CAGR according to Mordor Intelligence estimates.
The competitive edge in 2026 is no longer about who can hire the most people in Manila or Guadalajara. It is about who can most effectively integrate AI to lower the cost per resolution for global brands. The industry is exiting the “labor arbitrage” era and entering the “intelligence arbitrage” era.
How Incumbents Are Responding
| Company | Strategy | AI Revenue | Key Metric |
|---|---|---|---|
| TaskUs | Actively cannibalizing BPO to build AI moat | AI Services growing 60.8% YoY | 21.2% EBITDA margin (highest in sector) |
| Concentrix | Full AI pivot across operations | Significant investment | 12.3% EBITDA margin |
| Genpact | Analytics + AI process transformation | Embedded in delivery | 17.7% EBITDA margin |
| Accenture | Enterprise AI consulting + delivery | Multi-billion AI practice | Premium pricing |
TaskUs is the most instructive case study. The company is deliberately sacrificing short-term revenue growth — projecting only 3.5% growth in 2026 (down from 19% in 2025) — because increased automation at its largest client is displacing human-performed tasks. It is dedicating $25M+ in 2026 to AI transformation, betting that higher margins from AI services will more than compensate. Its 2026 revenue guidance is $1.21-$1.24B, and it launched an Agentic AI Consulting Practice to lead businesses through AI adoption.
The SMB Gap
While enterprise BPOs race to adopt AI, the SMB market remains underserved:
- 43% of SMBs cite lack of expertise as their primary AI adoption barrier
- 38% struggle with integration challenges
- 82% of businesses under 5 employees believe AI is not applicable to them
- 51% of business leaders identify insufficient AI knowledge at management/board level
Yet demand is real: SMB AI investment reached 57% in 2025, up from 36% in 2023 — a 58% rise in two years. They want AI. They just cannot build it themselves.
This is Trilled’s opening. The SMB market needs a trusted partner who can deliver both the human support they understand today and the AI capabilities they need tomorrow.
2. The Hybrid Human+AI Model
The Economics
Current cost benchmarks tell a clear story:
| Metric | AI Agent | Human (U.S.) | Human (Nearshore Mexico) |
|---|---|---|---|
| Cost per resolution | $0.50-$1.25 | $6.00-$8.00 | $2.00-$4.00 |
| Hourly cost | N/A | $28-$42/hr | $10-$18/hr |
| Availability | 24/7/365 | Shift-dependent | Shift-dependent |
| After-hours premium | None | 150-200% base | 150-200% base |
| Resolution rate (routine) | 70-85% | 95%+ | 95%+ |
| Complex issue handling | Poor | Excellent | Good-Excellent |
Critical insight from Gartner (January 2026): By 2030, GenAI cost per resolution will exceed $3, surpassing many B2C offshore human agents. Rising data center costs, AI vendor pricing shifts from subsidized to profitable, and increasingly complex use cases will drive this. Additionally, by 2027, half of companies that cut customer service staff due to AI will rehire people for similar functions.
This validates the hybrid model. Pure AI replacement is not the endgame — intelligent orchestration of human and AI capabilities is.
Optimal Ratio and Handoff Design
Mature deployments show AI can handle 70-85% of routine inquiries autonomously, with resolution rates reaching 81%. The remaining 15-30% require human intervention for:
- Complex multi-step issues
- Emotional/sensitive situations
- Edge cases outside training data
- Regulatory-required human interaction
- High-value customer retention scenarios
The handoff architecture:
- AI first: Every interaction starts with the AI agent
- Confidence scoring: AI evaluates its confidence in resolving the issue (threshold: typically 85%)
- Seamless escalation: Below-threshold or sentiment-flagged conversations route to human agents with full context summary
- Human resolution: Agent resolves with AI copilot assistance (suggested responses, knowledge retrieval)
- Feedback loop: Human resolution data trains the AI to handle similar cases in the future
The optimal staffing ratio evolves over time: - Month 1-3: 90% human / 10% AI (learning phase) - Month 4-6: 70% human / 30% AI (copilot phase) - Month 7-12: 40% human / 60% AI (agent phase) - Month 12+: 20% human / 80% AI (mature phase)
3. Revenue Model Evolution
The Pricing Transition
The BPO industry is moving through three distinct pricing phases:
Phase 1: Per-Seat / Per-Hour (Traditional) - Familiar to clients, easy to sell - Trilled starting point: $12-$20/hr for nearshore Mexico agents - Margin: 25-35% after fully loaded costs - Risk: Commoditized, race to the bottom
Phase 2: Hybrid Retainer + Performance Three revenue components emerge: - Human infrastructure retainer: $20,000-$50,000/month for AI trainers and escalation specialists - Technology platform fee: $4,000-$8,000/month for AI agent infrastructure - Performance incentives: Tied to automation rate, CSAT, and resolution metrics - Margin: 35-50%
Phase 3: Outcome-Based / Per-Resolution - $1-$5 per resolved interaction depending on complexity - Volume-based with quality guarantees - Margin: 50-70% as AI handles majority of volume - Industry direction per Andreessen Horowitz and enterprise CX procurement trends
Avoiding Revenue Cannibalization
The critical strategic challenge: how does Trilled avoid destroying its own revenue as AI replaces headcount?
TaskUs’s playbook provides the answer: 1. Price on outcomes, not inputs. When the client pays per resolution, it does not matter whether a human or AI resolves it — Trilled captures the margin either way 2. Expand scope. As AI handles tier 1, redeploy humans to higher-value services (QA, training, analytics, account management) 3. Sell AI as a product. The AI agent becomes a SaaS offering with recurring revenue, distinct from headcount 4. Land and expand. Start with one process, prove ROI, then offer AI across additional workflows (billing, onboarding, scheduling)
The margin math works in Trilled’s favor: if a human agent costs $1,800/month fully loaded in Mexico and handles 400 resolutions, that is $4.50/resolution. An AI agent handling 2,000 resolutions at $0.15 in token costs produces $0.15/resolution — a 30x margin improvement. Even at $2/resolution pricing, Trilled captures $1.85 margin per interaction vs. the $1.50 margin on human delivery.
4. Client Acquisition Strategy
Target Verticals
Based on AI adoption barriers, process complexity, and BPO readiness:
| Vertical | Why It Fits | AI Opportunity | Typical Contract |
|---|---|---|---|
| E-commerce / D2C | High ticket volume, repetitive queries (order status, returns) | 80%+ automatable | $8-15K/mo |
| SaaS / Tech | Technical support, onboarding, billing | 60-70% automatable | $12-25K/mo |
| Healthcare admin | Scheduling, insurance verification, patient intake | 50-60% automatable (regulated) | $15-30K/mo |
| Real estate / PropTech | Tenant inquiries, maintenance requests, showing scheduling | 70-80% automatable | $5-12K/mo |
| Financial services | Account inquiries, compliance docs, onboarding | 40-50% automatable (highly regulated) | $20-40K/mo |
| Manufacturing | Order processing, supplier communication, inventory inquiries | 60-70% automatable | $10-20K/mo |
The Sales Pitch
The pitch that converts is a Trojan horse strategy — sell the familiar, deliver the future:
“We provide dedicated, bilingual support teams in Mexico City at 40-60% lower cost than U.S. hires. Same timezone, same culture, no management overhead. And as we learn your business, we’ll progressively automate your most repetitive processes with AI — at no additional cost to you. You’ll see your cost per resolution drop while quality goes up.”
Key selling points for SMBs: 1. No AI risk for the client. They are buying human support. AI is Trilled’s internal efficiency play. 2. Timezone alignment. Mexico City is CST — same as Chicago, Dallas, Houston. 3. Cultural proximity. Mexican agents understand U.S. business norms. 4. Try before you buy. Start with 2-3 agents, scale based on results. 5. Built-in AI upside. As AI kicks in, the client sees declining cost per resolution without doing anything.
5. Architecture for Customer Support AI Agents
Core Architecture Pattern: Agentic RAG
The industry has moved beyond simple RAG pipelines. The current best practice is Agentic RAG — a system where the LLM acts as a reasoning engine with autonomy to self-correct, rewrite queries, perform multi-step reasoning, and use tools.
┌─────────────────────────────────────────────────┐
│ Orchestrator │
│ (LangGraph / Claude SDK) │
├─────────┬──────────┬──────────┬─────────────────┤
│ Router │ Retriever│ Resolver │ Escalation │
│ Agent │ Agent │ Agent │ Agent │
├─────────┴──────────┴──────────┴─────────────────┤
│ Tool Layer (MCP Servers) │
│ ┌─────────┐ ┌──────────┐ ┌───────────────────┐ │
│ │ Vector │ │ CRM / │ │ Ticketing System │ │
│ │ DB │ │ Client DB│ │ (Zendesk, etc.) │ │
│ └─────────┘ └──────────┘ └───────────────────┘ │
├─────────────────────────────────────────────────┤
│ Guardrails & Observability │
│ PII Detection | Hallucination Check | Logging │
└─────────────────────────────────────────────────┘
Multi-Agent Design
For customer support, a multi-agent architecture outperforms a single monolithic agent:
- Router Agent — Classifies intent, determines complexity, routes to appropriate handler
- Knowledge Retrieval Agent — Searches vector DB, knowledge base, past tickets for relevant context
- Resolution Agent — Generates the actual response using retrieved context and conversation history
- QA/Guardrail Agent — Validates response for accuracy, tone, PII, hallucination before delivery
- Escalation Agent — Determines when human handoff is needed, prepares context summary
Key Technical Decisions
- Stateful conversations: Use LangGraph’s built-in checkpointing for session continuity across interactions
- Tool use over fine-tuning: Equip agents with tools (API calls, DB queries, calculators) rather than trying to bake all knowledge into the model
- Streaming responses: Enable real-time streaming for chat interfaces to reduce perceived latency
- Async processing: Handle background tasks (ticket updates, CRM writes, email notifications) asynchronously
6. Knowledge Management and Ingestion
The Knowledge Ingestion Pipeline
Converting a client’s institutional knowledge into AI-usable form is where Trilled adds the most value. This is the moat — it requires human understanding to do well.
Ingestion hierarchy (priority order):
- Existing documentation — Help articles, FAQs, product docs, SOPs
- Historical tickets — Past support interactions with resolutions (gold mine)
- Internal wikis — Confluence, Notion, Google Docs
- Tribal knowledge — Captured through human agent observation and interviews
- Product/system behavior — API responses, error messages, UI flows
Technical Implementation
Chunking strategy: - Semantic chunking (not fixed-size) — split on logical boundaries (sections, paragraphs, FAQ pairs) - Optimal chunk size: 256-512 tokens with 50-token overlap - Metadata enrichment: tag each chunk with source, category, date, confidence
Embedding and retrieval: - Embedding models: OpenAI
text-embedding-3-large or open-source
nomic-embed-text for cost efficiency - Vector database:
Pinecone (managed, scales easily), Qdrant (self-hosted option), or
pgvector for simplicity - Hybrid retrieval: Combine vector similarity
search with BM25 keyword search for best results - Re-ranking: Use a
cross-encoder re-ranker (Cohere Rerank or open-source) to improve
precision
Multi-modal retrieval: - Vector search for semantic similarity - Knowledge graph traversal for relationship-based queries - Keyword search for exact match requirements (order numbers, product SKUs) - Metadata filtering for recency and category relevance
Key principle: Start with the content that supports the most frequent, highest-impact questions. Do not try to ingest everything at once. Companies report 40-60% faster resolution times when support agents have semantic access to their knowledge base.
7. Conversation Design and Quality
Design Principles
- Start helpful, not robotic. The AI should sound like a competent human agent, not a chatbot
- Acknowledge before solving. Mirror the customer’s concern before jumping to resolution
- Progressive disclosure. Give the most likely answer first, offer detail if needed
- Know when to stop. If the AI is not confident, escalate — do not guess
Escalation Logic
Escalation triggers (any one sufficient): - Confidence score below 85% on the proposed resolution - Negative sentiment detected (anger, frustration, threat) - Customer explicitly requests a human - Issue involves billing disputes above a threshold - Three consecutive failed resolution attempts - PII-sensitive operations (account changes, refunds) - Regulatory-required human interaction
Quality Metrics
| Metric | Target | Measurement |
|---|---|---|
| First Contact Resolution (FCR) | >75% | Resolved without escalation or follow-up |
| Customer Satisfaction (CSAT) | >4.2/5 | Post-interaction survey |
| Average Handle Time (AHT) | <3 min AI / <8 min human | Time from first message to resolution |
| Escalation Rate | <20% | Interactions requiring human handoff |
| Hallucination Rate | <1% | QA-flagged factually incorrect responses |
| Containment Rate | >80% | Resolved entirely by AI |
AI Conversation Summarization
When escalating to a human agent, the AI generates a structured handoff: - Issue summary: One-sentence description - Customer sentiment: Positive / Neutral / Frustrated / Angry - Actions taken: What the AI already tried - Relevant context: Order details, account info, past interactions - Suggested resolution: What the human should try next
8. Integration Patterns
The MCP Standard
The Model Context Protocol (MCP), introduced by Anthropic in November 2024, has become the industry standard for connecting AI agents to external systems. Adoption has been explosive: - November 2024: ~2M monthly SDK downloads - April 2025: 22M (OpenAI adoption) - July 2025: 45M (Microsoft integration) - November 2025: 68M (AWS Bedrock support) - March 2026: 97M monthly downloads
Gartner projects 40% of enterprise applications will integrate with task-specific AI agents by end of 2026, up from <5% previously.
Integration Map for Trilled
| System | Integration Method | Purpose |
|---|---|---|
| Zendesk | MCP Server + REST API | Ticket management, knowledge base |
| Intercom | MCP + Webhooks | Live chat, Fin Tasks integration |
| Salesforce | Agentforce MCP connectors | CRM, customer data, case management |
| Slack | MCP Server + Bot API | Internal communication, notifications |
| HubSpot | REST API + Webhooks | SMB CRM, marketing automation |
| Shopify | REST API + GraphQL | E-commerce order/product data |
| Google Workspace | MCP tools | Email, calendar, docs |
| Twilio | REST API + Webhooks | Voice/SMS channel |
| Stripe | REST API | Payment and billing queries |
Middleware Architecture
For clients with multiple systems, Trilled should build a unified middleware layer:
- n8n or Temporal for workflow orchestration
- MCP servers per integration (one per tool/system)
- Event bus (Redis Streams or Kafka for scale) for real-time data flow
- API gateway for rate limiting, auth, and routing
9. Data Access & Tooling Strategy
The Core Problem
You cannot build AI agents without access to the client’s support data, knowledge base, and internal systems. This is the single biggest bottleneck in the BPO-to-AI transition. If Trilled is selling BPO services today with the strategic intent of building AI on top of that data tomorrow, the tooling decisions made during client onboarding will determine whether the AI play is even possible.
The Pitch Problem
There is a tension between selling traditional BPO services and planning to build AI on the client’s data. The framing matters. “We need API access to your systems” sounds very different from a traditional BPO arrangement. The solution is to make programmatic access a natural part of how Trilled delivers the BPO service itself — not something bolted on later.
Recommended Tooling Stack
As part of onboarding, require (or strongly recommend) that clients use tools with robust APIs that Trilled can access programmatically:
| Function | Recommended Tools | Why |
|---|---|---|
| Ticketing | Zendesk, Intercom, Freshdesk | Mature APIs, webhook support, export capabilities. Every ticket becomes training data. |
| Knowledge Base | Notion, Confluence, shared Google Drive | Structured content that can be ingested into RAG pipelines. API access enables real-time sync. |
| Communication | Slack (with MCP access), shared channels | Captures tribal knowledge, escalation patterns, and real-time context. MCP integration is native. |
| CRM | HubSpot or Salesforce | Customer context enriches AI responses. API access enables automated data pulls. |
Handling Existing Client Tools
Not every client will use the recommended stack. The strategy varies by situation:
- Client has API-friendly tools: Integrate directly. Build an MCP server for their stack and move on.
- Client has legacy or closed tools: Propose migration as part of “modernizing their support operations.” This is a legitimate BPO value-add — you are improving their support infrastructure, which happens to also enable AI. Frame it as: “To deliver the best service, we need your team and ours working in tools that support real-time collaboration and reporting.”
- Client refuses to change: Accept it and work with what they have. Manual data export is a fallback. Not every client will be an AI candidate on day one.
Data Access Agreement
Include in the Master Services Agreement (MSA) a clause granting Trilled access to support data for “quality assurance, training, and process improvement.” This language covers AI training without explicitly saying “we are going to automate your people.” It is standard in BPO contracts — every major BPO includes similar language for workforce optimization and quality management.
Key MSA provisions: - Right to access and analyze all support interaction data - Right to use anonymized/aggregated data for internal process improvement - Data retention and handling aligned with client’s privacy requirements - Clear data ownership: client owns their data, Trilled has a license to use it for service delivery and improvement
The Transparent Approach
Some clients will be excited about AI from day one. For those, skip the Trojan horse and be direct:
“We’ll handle your support with a dedicated team in Mexico City. Over time, we’ll introduce AI that makes it faster and cheaper. You’ll see the savings in your invoice. The more data we have, the better the AI gets — so we need full access to your ticketing system and knowledge base.”
This approach works especially well with tech-savvy founders and SMBs that are already experimenting with AI but lack the bandwidth to build it themselves.
MCP as a Strategic Advantage
The Model Context Protocol is not just a technical convenience — it is a strategic moat. By building MCP servers for every client’s tool stack, Trilled creates:
- Reusable integrations: An MCP server for Zendesk works for every Zendesk client. Build once, deploy many.
- Real-time access: AI agents query live data (order status, account info, ticket history) without batch exports or data warehousing.
- Portability: MCP is model-agnostic. If Trilled switches from Claude to GPT or an open-source model, the integrations carry over.
- Speed: New client onboarding goes from weeks to days once the MCP server library is built out.
Data Pipeline Architecture
Set up continuous data ingestion from day one of every client engagement:
- Conversation logs: Every support interaction (chat, email, phone transcript) flows into a structured data store
- Resolution patterns: Tag each ticket with the resolution type, steps taken, and outcome
- FAQ mining: Automatically identify the most common questions and cluster them for RAG ingestion
- Knowledge base sync: Nightly or real-time sync of client knowledge base content into the vector database
- Performance signals: CSAT scores, resolution times, and escalation reasons feed back into the training loop
The pipeline should be automated and client-agnostic — the same architecture serves every client, with per-client isolation at the data layer.
10. Workforce Career Progression Model
Why This Matters
The AI-first BPO model only works if the humans in it see a future. High turnover in BPO is already a major cost driver — if agents believe AI is coming for their jobs with no alternative, retention will collapse. A clear career progression model solves three problems at once: it retains top talent, it creates the human-in-the-loop workforce that makes AI better, and it gives Trilled a scalable growth engine.
This model is also critical for Mexico labor law compliance. Documented performance management processes and clear promotion criteria protect Trilled in the event of terminations.
Tier 1 — Entry (Frontline Support Agents)
New hires join a client project as frontline support agents.
What they do: - Handle tickets across all channels (chat, email, phone) - Learn the client’s business, products, and SOPs - Follow established resolution playbooks - Use AI copilot tools to assist with responses
How they’re measured: - Resolution time - Customer satisfaction (CSAT) scores - Accuracy (QA review scores) - Ticket volume handled - AI copilot adoption rate (are they using the tools?)
What they don’t know: Every interaction they handle is training data. Their resolutions, corrections, and escalation decisions are feeding the AI that will eventually handle many of their routine tasks. This is not deceptive — it is the nature of the system. The best agents produce the best training data.
Typical duration: 3-6 months before promotion eligibility.
Tier 2 — Promoted Performers (Escalation + QA)
Top performers from Tier 1 advance into one of two specialized tracks:
Track A — Escalation Specialists: - Handle complex tickets that AI cannot resolve: multi-step issues, angry customers, edge cases, policy exceptions - Serve as the human backstop in the hybrid model - Their resolution quality sets the ceiling for what AI can eventually learn - Higher pay, higher responsibility, and genuine expertise development
Track B — AI Trainers / QA: - Review AI agent responses and flag errors - Write training examples and few-shot prompts - Refine system prompts based on failure analysis - Update and curate client knowledge bases - Serve as the human-in-the-loop for the AI improvement cycle
Why these roles are secure: Both tracks become MORE valuable as AI scales, not less. Escalation specialists handle the cases AI cannot — and those cases are the hardest, most nuanced, most human. AI trainers directly improve the system. Neither role is automatable.
Compensation: 20-40% above Tier 1, plus performance bonuses tied to AI quality metrics (for trainers) or complex resolution CSAT (for escalation).
Tier 3 — Redeployed (Rotation to New Clients)
Agents who performed adequately at Tier 1 but did not make the cut for Tier 2 promotion enter the rotation pool.
How rotation works: - When AI takes over enough of a client’s routine workflow, some Tier 1 agents are no longer needed on that account - These agents rotate to the next new client project that Trilled onboards - They bring institutional knowledge of “how BPO works at Trilled” — SOPs, tools, AI copilot usage — to the new project - They bootstrap the new engagement faster than a raw hire would
This is the flywheel: As AI matures on one client, freed-up agents seed the next client engagement. Trilled grows the client base without proportionally growing headcount. Net headcount still grows (because new clients are always being added), but cost-per-client drops over time.
Tier 4 — Exited (Performance Management)
Low performers are let go through documented performance management.
- AI-generated performance data (ticket quality scores, CSAT ratings, QA review results) provides objective justification
- Performance improvement plans (PIPs) are standard — typically 30-60 days
- This is natural attrition, not mass layoffs driven by automation
- Mexico labor law (Ley Federal del Trabajo) requires documented cause for termination without full severance — the AI-powered metrics system provides this documentation
The Math
For a typical client engagement:
| Phase | Agents | Roles | Timeline |
|---|---|---|---|
| Onboarding | 8 | All Tier 1 frontline | Month 1-3 |
| AI Copilot | 6 | 6 Tier 1 (AI-assisted, higher throughput) | Month 4-6 |
| AI Agent | 3 | 2 Escalation Specialists, 1 AI Trainer | Month 7-12 |
| Mature | 2 | 1 Escalation Specialist (shared), 1 AI Trainer (shared) | Month 12+ |
The other 5-6 agents either: - Promoted to Tier 2 on this or another client (best performers) - Rotated to the next new client project (adequate performers) - Exited through performance management (low performers)
Net headcount grows because Trilled keeps adding clients — but revenue-per-head increases dramatically.
Why This Model Works
- Retention: Good agents see a career path — from frontline to specialist to trainer — not a dead-end job waiting for the AI axe
- AI quality: The best humans become AI trainers, creating a virtuous cycle where top performers improve the system that handles routine work
- Scalability: Client base grows faster than headcount. Each new client requires fewer humans because the AI platform gets better with each deployment.
- Client optics: Trilled is “upskilling” its workforce and investing in career development — not replacing workers with robots
- Compliance: Documented performance metrics, clear promotion criteria, and structured PIPs satisfy Mexico labor law requirements for terminations
11. Training and Continuous Improvement
The Feedback Loop
The key advantage of the BPO + AI model: human agents generate training data every day.
Customer Interaction
│
┌────┴────┐
│ AI │ ──→ Resolved? ──→ Log as positive example
│ Agent │
└────┬────┘
│ Escalated
┌────┴────┐
│ Human │ ──→ Resolution captured as training data
│ Agent │ (ideal response for this scenario)
└─────────┘
Improvement Methods (Priority Order)
1. Prompt engineering (fastest, cheapest) - Iterative system prompt refinement based on failure analysis - Few-shot examples from best human agent responses - Most problems can be solved with better prompts — reach for this first
2. RAG knowledge base updates (days) - Add new FAQ entries, product updates, policy changes - Ingest recent ticket resolutions as new knowledge - Re-embed and re-index as content changes
3. Fine-tuning with DPO (weeks) - Collect preference pairs: AI response vs. human-corrected response - Direct Preference Optimization (DPO) is simpler and more stable than RLHF - DPO reframes preference learning as a classification problem — no reward model needed - Best for: consistent tone, domain-specific language, output format compliance - Use when: processing 50,000+ responses/day and even 2% error rate causes problems
4. Supervised fine-tuning (SFT) + DPO pipeline (months) - SFT provides a better starting point; DPO refines based on preferences - Only justified at significant scale or for highly specialized domains - Consider only after prompt engineering and RAG optimization are exhausted
Data Flywheel
Every client engagement generates data that makes Trilled’s AI better: - Cross-client learning: Patterns from e-commerce client A improve the agent for e-commerce client B (with appropriate data isolation) - Agent performance benchmarking: Compare AI vs. human resolution quality to identify improvement opportunities - Failure analysis: Weekly review of escalated/failed AI interactions to identify systematic gaps
12. Safety, Compliance, and Guardrails
Essential Guardrails
Guardrails operate at runtime, blocking or transforming unsafe content within 200-300ms latency budgets:
- Hallucination detection — Validate AI responses against retrieved source documents. Tools: Patronus AI (Lynx model outperforms GPT-4 on hallucination benchmarks), Galileo AI
- PII detection and redaction — Scan inputs and outputs for personal data. Auto-redact SSNs, credit cards, addresses before processing
- Prompt injection prevention — Detect and block attempts to manipulate the AI through adversarial inputs
- Topic guardrails — Prevent AI from discussing topics outside its scope (politics, medical advice, legal advice unless qualified)
- Tone and brand compliance — Ensure responses match client brand guidelines
Regulatory Landscape
| Regulation | Scope | Key Requirements |
|---|---|---|
| GDPR | EU data subjects | Data minimization, right to erasure, consent, DPO |
| CCPA/CPRA | California residents | Opt-out rights, data access, deletion requests |
| EU AI Act (effective 2025) | EU market | Risk classification, high-risk system requirements. Fines up to EUR 35M or 7% global revenue |
| Mexico’s LFPDPPP | Mexican data subjects | Consent, ARCO rights, privacy notices |
| HIPAA | U.S. healthcare data | BAA requirements, encryption, access controls |
| PCI-DSS | Payment card data | No storage of card data, encryption in transit |
Compliance Certifications to Target
For enterprise credibility, Trilled should pursue: - SOC 2 Type II — Table stakes for handling client data - ISO 27001 — Information security management - ISO 42001 — AI management system standard (emerging requirement)
Human-in-the-Loop Requirements
Gartner predicts regulatory pressure will mandate easy access to human agents, with customers being encouraged to request a human by default. Trilled’s hybrid model is inherently compliant with this trend — humans are always available.
13. AI for Recruiting Operations
The AI Recruiting Stack
Trilled needs to recruit efficiently to staff BPO operations. The 2026 AI recruiting landscape:
| Tool | Capability | Best For | Pricing Model |
|---|---|---|---|
| Paradox (Olivia) | Conversational AI for screening, scheduling, onboarding | High-volume, hourly roles | Per-hire |
| HireVue | Video interviewing + AI assessment | Behavioral evaluation | Per-assessment |
| Fetcher | AI-powered sourcing and outreach | Passive candidate sourcing | Subscription |
| Findem | Autonomous sourcing agent | Full pipeline automation | Subscription |
| Juicebox | AI sourcing with autonomous outreach | Tech roles | Per-seat |
2026 trend: Autonomous AI agents now operate independently — sourcing candidates, sending outreach, scheduling interviews, and delivering shortlists without manual intervention. AI tools automate 60-70% of recruiter time on repetitive tasks.
Build vs. Buy Recommendation
| Function | Recommendation | Rationale |
|---|---|---|
| Resume screening | Buy (Paradox) | Commodity capability, not worth custom build |
| Interview scheduling | Buy (Paradox/Calendly AI) | Solved problem |
| Candidate sourcing | Buy (Fetcher/LinkedIn Recruiter) | Network access required |
| Skills assessment | Build custom | Domain-specific for BPO roles (typing speed, English proficiency, customer empathy) |
| Cultural fit screening | Build custom | Trilled-specific values and work style |
| Onboarding automation | Build custom | Integrates with training pipeline |
14. Mexico-Specific Recruiting
Talent Market
Mexico produces 1,009,855 university graduates annually (2024-2025 cycle), providing access to bilingual customer service representatives, data analysts, software developers, and administrative professionals.
Regional talent hubs: - Mexico City — Financial services, professional services, largest talent pool - Guadalajara — Technology hub, strong software development ecosystem - Monterrey — Industrial/business center, strong in tech and manufacturing - Tijuana — BPO specialization, bilingual customer support
Sourcing Channels
| Channel | Best For | Cost |
|---|---|---|
| Professional/tech roles | Premium subscription | |
| OCC Mundial | General professional roles | Per-posting |
| Computrabajo | Entry-level, high volume | Per-posting |
| Indeed Mexico | Broad reach | Pay-per-click |
| University partnerships | Pipeline building | Relationship-based |
| Referral programs | Highest quality hires | Bonus per hire |
Labor Law Compliance (Ley Federal del Trabajo)
Critical requirements: - Aguinaldo: Minimum 15 days salary, paid before December 20 - Vacation: Minimum 12 days after first year (increased per 2023 reform) - Vacation premium: 25% of vacation pay - Profit sharing (PTU): 10% of pre-tax profits distributed to employees - Social security (IMSS): Mandatory employer contributions - Severance: 3 months salary + 20 days per year worked
Retention Strategy
In BPO environments, high turnover destroys service quality. The top 5 desired perks in Mexico: 1. Savings fund (fondo de ahorro) 2. Paid training and development 3. Performance bonuses 4. Family benefits (medical for dependents) 5. Paid overtime (not just comp time)
Salary benchmarks for BPO roles in CDMX: - Entry-level bilingual agent: $12,000-$18,000 MXN/month - Senior agent / team lead: $20,000-$30,000 MXN/month - QA / Training specialist: $18,000-$25,000 MXN/month - Operations manager: $35,000-$50,000 MXN/month
15. LLM Selection Guide
Model Comparison for Customer Support (April 2026)
| Model | Input/Output Cost (per 1M tokens) | Speed (tokens/sec) | Best Use Case |
|---|---|---|---|
| Claude Sonnet 4.6 | $3 / $15 | ~77 | Complex reasoning, nuanced responses |
| Claude Haiku | $0.25 / $1.25 | ~150 | High-volume routing, simple queries |
| GPT-4o | $2.50 / $10 | ~100 | General purpose, good balance |
| GPT-4o mini | $0.15 / $0.60 | ~180 | Cost-sensitive high volume |
| Gemini 2.5 Flash | $0.10 / $0.40 | ~250 | Lowest cost, speed-critical |
| Llama 4 (self-hosted) | Infrastructure cost only | Varies | Data sovereignty, regulated industries |
| DeepSeek V3.2 | $0.28 / $0.42 | ~120 | Budget alternative, strong performance |
Recommended Model Strategy
Tiered approach by interaction complexity:
- Router/classifier: GPT-4o mini or Gemini Flash ($0.10-$0.15/1M tokens) — classify intent, route to appropriate handler
- Tier 1 resolution (routine): Claude Haiku or GPT-4o mini — FAQ answers, order status, simple troubleshooting
- Tier 2 resolution (complex): Claude Sonnet 4.6 or GPT-4o — multi-step reasoning, policy interpretation
- Quality assurance: Claude Sonnet — validate responses before delivery
Semantic caching: A support agent answering variations of the same 20 questions can cut costs by 60-80% through semantic caching (hash similar queries, return cached responses).
Self-hosting consideration: Fine-tuned Llama variants are viable for regulated industries where data cannot leave infrastructure. Quantized variants and deployment scripts (via Ollama, vLLM, or TGI) lower self-hosting barriers significantly.
16. Agent Framework Comparison
Detailed Framework Assessment
| Feature | LangGraph | CrewAI | Claude Agent SDK | AutoGen | Semantic Kernel |
|---|---|---|---|---|---|
| Production readiness | Highest | Medium | Medium-High | Medium | Medium-High |
| Learning curve | Medium (graph concepts) | Low (role-based DSL) | Low (tool-use) | High | Medium |
| State management | Built-in checkpointing + time travel | Sequential task output | MCP server state | Conversation-based | Plugin-based |
| Model flexibility | Any model | Any model | Claude only | Any model | Any model |
| MCP support | Good | Best (3 transport modes) | Native (in-process, zero latency) | Limited | Good |
| Observability | LangSmith (best-in-class) | Community tools | Basic | Limited | Azure Monitor |
| GitHub stars | 15K+ | 44.6K | Growing | 40K+ | 25K+ |
| Lines to start | ~60 | ~20 | ~30 | ~80 | ~50 |
| Best for | Complex stateful workflows | Rapid prototyping, multi-agent | Claude-native apps | Research, experimentation | Microsoft ecosystem |
Recommendation for Trilled
Phase 1-2: Start with CrewAI for rapid prototyping and validation. Get multi-agent workflows running in under an hour with minimal code. The role-based DSL maps naturally to support team roles (router, resolver, QA).
Phase 3-4: Migrate to LangGraph for production deployment. The most common industry migration path is CrewAI to LangGraph — prototype in CrewAI, validate the concept, then migrate by mapping each CrewAI agent to a LangGraph node. LangGraph’s graph-based architecture handles conditional branching, parallel execution, and complex state management essential for multi-tier support escalation.
MCP integration: Use MCP servers for all tool integrations from day one, regardless of framework choice. This ensures tool implementations are portable between frameworks.
17. Evaluation and Monitoring
Observability Stack
89% of organizations have implemented observability for their AI agents, with quality issues being the primary production barrier (32%).
Recommended tools:
| Tool | Function | Why |
|---|---|---|
| Braintrust | Comprehensive agent traces, automated evaluation, cost analytics | Best overall platform |
| Langfuse | Self-hosted LLM observability, prompt versioning, cost tracking | Privacy-first, self-hosted option |
| LangSmith | Trace viewing, evaluation, dataset management | Best if using LangGraph |
| Maxim AI | Simulation + evaluation + observability unified | Full lifecycle coverage |
Metrics Dashboard
Track daily: - Resolution rate: % of interactions fully resolved by AI - Escalation rate: % requiring human handoff - CSAT score: Post-interaction rating (target: >4.2/5) - Average handle time: AI vs. human - Cost per resolution: Actual token + infrastructure costs - Hallucination rate: QA-flagged incorrect responses - Latency: Time to first token, total response time
Track weekly: - One-Answer Success Rate: % resolved in single exchange - Containment rate: % staying within AI without human touch - Revenue impact: Support interactions that led to upsell/retention - Knowledge gap analysis: Topics where AI fails most frequently
18. Cost Modeling
Per-Client Unit Economics
Assumptions: Mid-size e-commerce client, 3,000 support tickets/month
Phase 1: Pure Human (Month 1-3)
| Item | Cost |
|---|---|
| 3 agents fully loaded (CDMX) | $5,400/mo ($1,800 each) |
| Team lead (fractional) | $1,500/mo |
| Tools & infrastructure | $500/mo |
| Total cost | $7,400/mo |
| Revenue (at $15/hr, 3 agents) | $10,800/mo |
| Margin | $3,400 (31%) |
Phase 2: AI Copilot (Month 4-6)
| Item | Cost |
|---|---|
| 2 agents fully loaded (1 reduced via productivity gain) | $3,600/mo |
| Team lead (fractional) | $1,500/mo |
| AI infrastructure (tokens, vector DB, hosting) | $1,500/mo |
| Tools & integrations | $500/mo |
| Total cost | $7,100/mo |
| Revenue (same contract) | $10,800/mo |
| Margin | $3,700 (34%) |
Phase 3: AI Agent (Month 7-12)
| Item | Cost |
|---|---|
| 1 agent (complex cases only) | $1,800/mo |
| AI infrastructure (higher volume) | $2,500/mo |
| Tools & integrations | $500/mo |
| Total cost | $4,800/mo |
| Revenue (renegotiated to $8K outcome-based) | $8,000/mo |
| Margin | $3,200 (40%) |
Phase 4: Mature AI (Month 12+)
| Item | Cost |
|---|---|
| 0.5 FTE (escalation specialist, shared) | $900/mo |
| AI infrastructure | $2,000/mo |
| Tools & integrations | $500/mo |
| Total cost | $3,400/mo |
| Revenue (outcome-based, lower per-unit but higher margin) | $6,000/mo |
| Margin | $2,600 (43%) |
Scaling Economics
| Metric | 5 Clients | 15 Clients | 50 Clients |
|---|---|---|---|
| Monthly revenue | $50K | $135K | $400K |
| Human headcount | 8 | 15 | 30 |
| AI infrastructure | $8K | $18K | $45K |
| Total costs | $30K | $72K | $185K |
| Monthly profit | $20K | $63K | $215K |
| Margin | 40% | 47% | 54% |
At 50 clients with mature AI deployment, Trilled hits $2.58M annual profit — exceeding the $200K USD/yr target by 12x. The model achieves the target at approximately 5-7 clients in Phase 3+.
Infrastructure Cost Breakdown
Monthly operational costs at moderate scale: - LLM API tokens: $1,000-$5,000/mo (depends on volume and model mix) - Vector database hosting: $100-$2,000/mo (Pinecone or Qdrant) - Embedding generation: Negligible at support volumes - Monitoring/observability: $500-$2,000/mo - Compute (API servers, workers): $500-$2,000/mo
19. Implementation Roadmap
Phase 1: BPO Foundation (Months 1-6)
Objective: Establish profitable BPO operations, learn client businesses deeply, build the data foundation for AI.
Actions: - [ ] Hire initial team: 5-8 bilingual agents, 1 team lead, 1 QA specialist - [ ] Set up operations: Office/coworking in CDMX, tools (Zendesk/Intercom, Slack, time tracking) - [ ] Win first 3-5 clients via per-seat pricing ($12-$18/hr) - [ ] Implement rigorous documentation: Every SOP, every edge case, every resolution path - [ ] Build knowledge base architecture: Obsidian/Notion vault per client with structured categories - [ ] Begin logging all interactions in structured format (training data for Phase 2) - [ ] Establish QA process: Weekly ticket reviews, CSAT tracking, agent coaching
Key hires: Operations manager, senior agents with BPO experience
Revenue target: $15-25K/mo Investment: $30-50K setup costs
Phase 2: AI Copilot (Months 4-9)
Objective: AI assists human agents — fastest path to value without client-facing risk.
Actions: - [ ] Deploy AI copilot for agents: suggested responses, auto-summarization, knowledge retrieval - [ ] Build RAG pipeline: Ingest client knowledge bases, historical tickets, SOPs - [ ] Set up vector database and embedding pipeline (Pinecone + OpenAI embeddings) - [ ] Implement auto-ticket classification and routing - [ ] Deploy conversation summarization for shift handoffs - [ ] Build internal dashboard: agent productivity metrics, AI suggestion acceptance rate - [ ] Begin A/B testing: AI-suggested vs. human-only responses (quality comparison) - [ ] Select and deploy observability (Langfuse self-hosted for cost efficiency)
Key hires: AI engineer (can be remote), data engineer (part-time)
Productivity gain target: 30-40% more tickets per agent Investment: $15-25K for AI infrastructure and engineering
Phase 3: AI Agent (Months 8-15)
Objective: AI handles tier 1 autonomously, humans handle complex cases.
Actions: - [ ] Deploy client-facing AI agent for tier 1 (with client approval and careful rollout) - [ ] Start with lowest-risk client and highest-volume, most-routine ticket types - [ ] Implement full escalation logic with sentiment detection - [ ] Build guardrail layer: hallucination detection, PII redaction, topic boundaries - [ ] Migrate first client to outcome-based pricing (pilot) - [ ] Redeploy freed human agents to new clients or higher-value work - [ ] Implement DPO fine-tuning pipeline using human agent corrections - [ ] Build client-facing analytics dashboard showing AI performance metrics - [ ] Begin SOC 2 compliance process
Key hires: Senior AI engineer, customer success manager
Target: 60-70% AI resolution rate, 40%+ margins Investment: $25-40K engineering + compliance
Phase 4: Platform (Months 12-24)
Objective: Package AI + BPO as a repeatable product. Scale beyond individual client implementations.
Actions: - [ ] Build multi-tenant AI platform: each client gets isolated knowledge base, custom agent config - [ ] Create self-service onboarding: client uploads docs, AI agent is provisioned automatically - [ ] Develop white-label option for other BPOs to use Trilled’s AI infrastructure - [ ] Build marketplace of pre-built integrations (Shopify, HubSpot, Zendesk, etc.) - [ ] Launch AI-only tier: clients who do not need human agents can subscribe to AI agent as SaaS - [ ] Expand to additional verticals and geographies - [ ] Consider raising capital for accelerated growth
Key hires: Product manager, 2-3 additional engineers, sales team
Target: 50+ clients, $400K+/mo revenue, 50%+ margins Investment: $100-200K (or fundraise)
20. Accelerated Roadmap (AI-Native Engineering)
Why the Original Timeline Is Too Conservative
Section 19 lays out a 24-month roadmap that assumes Trilled will hire human engineers to build the AI platform. That assumption is outdated. With Rune — an always-on Claude Code agent running on a dedicated Mac Pro — Trilled has an AI engineering team available today, at near-zero marginal cost. This changes the math on everything: timeline, hiring, investment, and sequencing.
The original roadmap treats BPO establishment and AI development as sequential phases. In reality, they can and should run in parallel. Every day that Trilled operates a BPO without building AI infrastructure is a day of training data going uncaptured and margin improvement left on the table.
What Changes
| Dimension | Original Roadmap | Accelerated Roadmap |
|---|---|---|
| Total timeline | 24 months to platform | 6-9 months to platform |
| First engineering hire | Month 4-9 (AI engineer) | Month 6-9 (Technical Account Manager) |
| Engineering cost (Year 1) | $80-120K USD | Near zero (Rune + API costs) |
| BPO and AI sequencing | Sequential — BPO first, then AI | Parallel — both from month 1 |
| First AI deployment | Month 4-6 (copilot) | Month 1-2 (internal on LINQ) |
| Breakeven timeline | Month 8-12 | Month 3-5 |
Revised Phase Plan
Month 1: Foundation Sprint - Build core AI agent framework: RAG pipeline, MCP server library, orchestration layer - Simultaneously: close next BPO client beyond LINQ - Deploy AI copilot internally on LINQ support tickets immediately — this is the proof of concept - Set up data pipeline to capture every LINQ interaction as structured training data - No human engineers needed. Rune builds, tests, and iterates on the framework.
Month 2-3: Internal Proof of Concept - Run AI copilot on live LINQ tickets. Measure everything: resolution time improvement, cost savings, agent productivity gains, AI suggestion acceptance rate. - Iterate on the system daily based on real failure modes — not theoretical ones - Build the internal dashboard showing AI performance metrics - Generate the case study: “Here is what happened when we deployed AI on a real client” - Begin building MCP servers for common client tools (Zendesk, Intercom, HubSpot)
Month 4-6: Sell with Data - Pitch new clients with real performance data from the LINQ deployment - The pitch changes from “we plan to introduce AI eventually” to “here are the actual numbers from our existing deployment — 40% faster resolution, 30% cost reduction, same CSAT scores” - Each new client onboards onto the same AI framework with their own isolated knowledge base - Rune builds per-client customizations (MCP servers, prompt configurations, knowledge ingestion) as each client signs
Month 6-9: Platform-ify - The framework hardens into a multi-tenant platform through real usage - Self-service onboarding: client uploads docs, AI agent is provisioned within days (not weeks) - First human technical hire: a Technical Account Manager who handles client-facing communication, not engineering. Rune handles the engineering. - Begin exploring outcome-based pricing with clients who have 3+ months of AI performance data
Month 9-12: Scale - Platform is stable, multi-client, and generating real margin improvement - Consider first engineering hire for 24/7 monitoring, on-call, and client-facing technical roles - Evaluate SOC 2 timeline based on enterprise client pipeline - Begin cross-client learning: patterns from Client A improve the agent for Client B
What Stays the Same
Not everything accelerates. Some things require real-world reps regardless of engineering speed:
- Knowledge ingestion quality still depends on understanding the client’s business deeply. AI can ingest documents, but knowing which edge cases matter requires human judgment and client interaction time.
- Client trust builds over months of reliable service, not days. Even with superior AI metrics, clients need to see sustained performance before moving to outcome-based pricing.
- Edge case discovery only happens through production volume. The long tail of weird customer issues that break AI agents cannot be anticipated — only observed and fixed.
- Regulatory and compliance work (SOC 2, data processing agreements) moves at the speed of auditors and lawyers, not engineers.
Investment Profile
The accelerated roadmap dramatically changes the cost structure:
| Cost Category | Original (Year 1) | Accelerated (Year 1) |
|---|---|---|
| Engineering salaries | $80-120K | $0 |
| AI infrastructure (APIs, hosting) | $15-25K | $20-30K (higher because deploying earlier) |
| Rune operating cost | N/A | ~$3-5K (API costs for Claude Code) |
| BPO setup (office, hiring, tools) | $30-50K | $30-50K (unchanged) |
| Total Year 1 investment | $125-195K | $53-85K |
Breakeven arrives faster because: 1. AI margin improvements start in month 2, not month 8 2. No engineering salary burn during the build phase 3. Real performance data closes new clients faster than a pitch deck
When to Hire Humans
The first human engineering hire is not an engineer — it is a Technical Account Manager (TAM). This person: - Manages client relationships during AI onboarding - Translates client needs into technical requirements that Rune implements - Monitors AI performance dashboards and escalates issues - Handles SOC 2 and compliance conversations with enterprise prospects
Timing: Month 6-9, when Trilled has 3-5 active clients and the TAM role becomes a bottleneck that AI cannot solve (because it requires human trust-building and in-person meetings).
The first pure engineering hire comes when Trilled needs: - 24/7 on-call coverage for production AI systems - Client-facing technical demos and architecture reviews - Infrastructure work that exceeds what Rune can handle autonomously (unlikely before 10+ clients)
21. Risks and Mitigations
| Risk | Severity | Likelihood | Mitigation |
|---|---|---|---|
| AI costs rise faster than expected (Gartner: >$3/resolution by 2030) | High | Medium | Hybrid model ensures humans remain available; tiered model selection reduces cost; self-hosting option via Llama |
| Client resistance to AI | Medium | High | Trojan horse approach — sell human, introduce AI gradually. Client never has to “buy AI” |
| Hallucination/quality failure | High | Medium | Multi-layer guardrails, human QA review of AI outputs, gradual rollout starting with low-risk queries |
| Data privacy breach | Critical | Low | SOC 2 compliance, encryption at rest/in transit, PII auto-redaction, data isolation per client |
| Employee resistance/turnover | Medium | Medium | Position AI as tool that makes agents’ jobs better (handles boring tickets). Upskilling path to AI trainer/QA roles |
| Competition from incumbents | Medium | High | Speed advantage — Trilled can move faster than TaskUs/Concentrix on SMB. Niche focus on specific verticals |
| Regulatory changes (EU AI Act, Mexico regulations) | Medium | Medium | Design for compliance from day one. Human-in-the-loop satisfies most requirements |
| Key client concentration risk | High | Medium | Diversify to 10+ clients within 12 months. No client should be >30% of revenue |
| Mexico peso appreciation | Low | Low | Contracts in USD. Peso costs are fixed; revenue is dollar-denominated |
| LLM vendor lock-in | Medium | Medium | MCP-based tool architecture is model-agnostic. Can switch between Claude, GPT, Llama |
Appendix: Recommended Technology Stack
Core Infrastructure
- LLM Provider: Anthropic (Claude) primary, OpenAI secondary, Ollama for local/development
- Agent Framework: CrewAI (prototype) → LangGraph (production)
- Vector Database: Pinecone (managed) or pgvector (cost-efficient)
- Orchestration: n8n or Temporal for workflow automation
- Observability: Langfuse (self-hosted) → Braintrust (scale)
Client-Facing
- Ticketing: Zendesk or Intercom (client’s choice)
- Chat Widget: Custom or Intercom Messenger
- Voice: Twilio
- CRM Integration: MCP servers per platform
Internal Operations
- Knowledge Management: Obsidian (internal), client-specific vaults
- Communication: Slack
- Project Management: Linear or Notion
- Recruiting: Paradox + custom screening
- HR/Payroll: Local Mexico payroll provider (Runa, Worky)
Security & Compliance
- Secrets Management: Doppler or AWS Secrets Manager
- PII Detection: Presidio (open-source) or Patronus AI
- Logging: Structured logging with PII redaction
- SOC 2: Vanta or Drata for compliance automation
Conclusion
The AI-first BPO model is not theoretical — it is being validated by public companies spending hundreds of millions on the transition. Trilled’s advantage is speed, focus, and positioning: a small, agile company in Mexico City that can move faster than incumbents, serve the underserved SMB market, and build the AI capabilities that turn a services business into a technology platform.
The path is clear: 1. Sell what SMBs already buy (human support teams) 2. Learn their businesses (knowledge capture is the moat) 3. Deploy AI progressively (copilot → agent → platform) 4. Capture expanding margins (40% → 50% → 60% as AI scales)
The $200K USD/yr profit target is achievable with 5-7 mature clients. The platform vision at 50+ clients produces $2.5M+ annually. The opportunity window is now — before incumbents figure out how to serve the SMB market and before AI-native startups build their own service layers.
Document prepared April 2026. Data and pricing subject to rapid change in the AI industry. Review and update quarterly.
[[trilled]] | [[projects/trilled]] | #AI #BPO #strategy