Innovatrix Infotech
How We Built an Agentic Workflow That Saves Our Clients 15+ Hours a Week cover
AI & LLM

How We Built an Agentic Workflow That Saves Our Clients 15+ Hours a Week

A laundry management client was spending 32+ hours a week manually answering WhatsApp queries. We built a three-agent workflow in n8n that now handles 78% of all queries autonomously, recovering 130+ hours per month. Here is the architecture, what broke the first time, and what actually made it work.

Rishabh SethiaRishabh SethiaFounder & CEO16 March 202613 min read2k words
#ai-automation#agentic-workflow#n8n#whatsapp-automation#case-study

A laundry management business was drowning in WhatsApp messages.

Not figuratively. Literally — 200+ customer messages per day, handled manually by a small team. Pickup scheduling, order status queries, complaint handling, pricing questions, custom service requests. The kind of repetitive, high-volume communication work that eats operational capacity alive.

When they came to us, their team was spending over 32 hours every week just responding to routine WhatsApp queries. That's almost a full-time employee, every week, doing work that produced zero strategic value.

We built them an agentic workflow that now handles the vast majority of that work autonomously. Within 60 days, their team had reclaimed 130+ hours per month of operational time.

Here's exactly how we did it — what we built, what broke the first time, and what made it actually work in production.


The Problem: 32 Hours a Week Answering the Same 12 Questions

Before we built anything, we mapped every incoming WhatsApp query over a two-week period. The result was predictable but clarifying: roughly 80% of all messages fell into 12 categories.

Pickup scheduling requests. Order status updates. Pricing for standard vs. premium service. Estimated delivery times. Item-specific handling questions (leather? silk? wedding dress?). Complaint escalations. Referral code inquiries. Reorder requests. Payment confirmation. Service area questions. Profile update requests. And the occasional general "hello, anyone there?" message.

The other 20% were genuinely complex: complaints with legal implications, novel service requests, items requiring individual assessment, upset customers who needed a human.

This 80/20 split is the foundational insight for any agentic workflow. If 80% of your work is structured, repeatable, and answerable from a known data set, that 80% is the automation target. The 20% that requires judgment, empathy, or novel reasoning? That stays human. That's not a failure of the system — it's the design.


The Solution Architecture: A Three-Agent WhatsApp System

We built the system in n8n, integrated with the WhatsApp Business API, and connected it to the client's existing order management database.

The architecture uses three agents:

Agent 1: Intent Classifier

Every incoming WhatsApp message is first processed by a classification agent. Its only job is to categorize the query into one of the known 12 categories, or flag it as "novel/complex." It also extracts key entities: customer phone number, order ID if mentioned, service type requested.

This agent runs in under 400ms on average. It never responds to the customer — it's purely an internal routing layer.

Agent 2: Knowledge + Response Agent

For any query that falls into the 12 known categories, the response agent handles the full conversation turn. It has access to:

  • The customer's order history and current status via API
  • A structured knowledge base of pricing, service areas, turnaround times, and policies
  • Response templates calibrated for the client's tone (friendly, professional, slightly informal — matching how their human team had been writing)

It generates a draft response, runs a self-check against the knowledge base to verify any factual claims (pickup timing, pricing figures), and then either sends the response or — if the self-check flags uncertainty — routes to the human queue.

Agent 3: Escalation Router

Any "novel/complex" flag from the classifier, any response that fails the self-check, and any message containing specific trigger keywords (complaint, legal, refund over a threshold, certain emotional indicators) gets routed to the human queue with full context: the original message, the customer's order history, and the agent's tentative response if one was drafted.

The human agent can approve the draft response (one click), edit it, or start a fresh reply. The AI did the research; the human makes the final call.

This is the human-in-the-loop pattern applied correctly: not every message requires approval, only the ones that carry real risk or uncertainty. The result is a system that's genuinely fast for routine work and genuinely safe for edge cases.


What Broke the First Time (This Is the Important Part)

The first version of the response agent had a problem we hadn't anticipated: it was too confident.

When a customer asked about a service we didn't offer — professional suit pressing, which wasn't in the knowledge base — the agent didn't say "I'm not sure about that." It confabulated a plausible-sounding answer based on its general knowledge of laundry services.

It told a customer we offered a service we didn't offer.

One message. The customer came in expecting the service. The client was embarrassed. We learned.

The fix was a combination of two changes:

Fix 1: Scope-bounded knowledge retrieval. The response agent can only cite information that exists in the structured knowledge base. It cannot generate answers from general training knowledge when no document in the knowledge base supports the claim. Full stop.

Fix 2: Explicit "I don't know" routing. If the agent cannot find a matching entry in the knowledge base with >85% confidence, it routes to the human queue with a flag: "Customer asked about: [topic]. No entry found in knowledge base. Requires human response."

This two-part fix eliminated the confabulation problem entirely. The human queue volume went up slightly in the short term — more "unknown" queries being flagged correctly — but the quality of automated responses increased dramatically. The client's team was only seeing genuinely hard questions, not being asked to fix AI-generated misinformation.

This is a pattern we now build into every knowledge-backed agent from day one. The lesson: an AI that says "I don't know" is not a failure. An AI that confidently makes things up is a liability.


The Results: 60 Days In

130+ hours per month reclaimed from manual WhatsApp handling. That's the headline number.

Behind it:

  • 78% of all queries now handled fully autonomously, start to finish, with zero human involvement
  • Average response time dropped from 2–4 hours (when a human was busy) to under 3 minutes
  • Human queue volume reduced from 200+ items/day to approximately 45 items/day — all of which are genuinely complex and require judgment
  • Customer satisfaction held steady through the transition (tracked via post-interaction satisfaction pings), with a slight uptick attributed to faster response times on routine queries
  • Zero confabulation incidents after the scope-bounding fix was deployed

The client's operations manager now spends her time on staff management, quality oversight, and business development — not answering "what time is my pickup?" for the fourteenth time on a Tuesday.


The Technical Stack (For Developers Who Want the Details)

The full system runs on:

  • n8n (self-hosted on AWS EC2) as the workflow orchestration layer
  • WhatsApp Business API via Meta's Cloud API for message ingestion and sending
  • Anthropic Claude Sonnet as the LLM backbone for both classification and response generation
  • PostgreSQL for the structured knowledge base (pricing, policies, service area data)
  • REST API integration with the client's order management system for real-time order status lookups
  • Slack webhook for human queue notifications — the team receives a Slack ping with full context for every escalated query

Total infrastructure cost: under $80/month. The LLM API cost is minimal at this query volume. The n8n instance runs on a t3.small EC2 instance.

The ROI math is straightforward. 130 hours/month at a conservative ₹200/hour blended labour cost = ₹26,000/month in recovered operational capacity. Monthly infrastructure cost: under ₹7,000. The system recovered its implementation cost within 6 weeks of deployment.

For a deeper look at how these workflows are architected, see our guide to building multi-agent workflows in n8n and the multi-agent systems explained post for the underlying architectural theory.


What This Pattern Applies To (Beyond Laundry)

The architecture — classifier → knowledge-backed response agent → escalation router — applies to any business with high inbound communication volume and a high proportion of repeatable query types.

We've built variants of this for:

  • D2C e-commerce order status and returns handling via WhatsApp and email, integrated with Shopify on the backend. If you're running a Shopify storefront and handling order queries manually, this is one of the highest-ROI automation investments available to you. See our Shopify development work for how the backend integration connects.
  • SaaS customer support tier-1 triage where the agent handles all FAQ-class queries and routes novel product issues to the engineering team
  • Internal IT helpdesk automation for a distributed team across time zones — the agent handles password resets, access requests, and known error resolutions 24/7 without human involvement

The key variable in all of them: the 80/20 split still holds. Map your query types before you build anything. If you can't show that at least 60–70% of your volume is repeatable and answerable from a knowledge base, the automation ROI math gets much harder to justify.


Want This Built for Your Business?

If your team is spending meaningful hours per week on repetitive communication work — customer support, order management, internal helpdesk, client status updates — an agentic workflow is likely the highest-ROI automation investment you can make right now.

We scope and price these as fixed-cost engagements. No surprise billing, no hourly overruns. See our AI automation services for how we structure these projects, and explore your use case with us if you want a realistic assessment of what automation can achieve for your specific volume and query mix.


Frequently Asked Questions

Written by

Rishabh Sethia

Rishabh Sethia

Founder & CEO

Rishabh Sethia is the founder and CEO of Innovatrix Infotech, a Kolkata-based digital engineering agency. He leads a team that delivers web development, mobile apps, Shopify stores, and AI automation for startups and SMBs across India and beyond.

Connect on LinkedIn
Get started

Ready to talk about your project?

Whether you have a clear brief or an idea on a napkin, we'd love to hear from you. Most projects start with a 30-minute call — no pressure, no sales pitch.

No upfront commitmentResponse within 24 hoursFixed-price quotes