Back to insights
ai-automationcustom-toolsllm-setuppillar

The Complete Guide to Building Custom AI Tools for Growth Teams

Everything growth teams need to know to ship a production AI tool — picking the right workflow, evaluating models, building context, setting up evals, and operating the system after it ships. Grounded in real production engagements, not vendor demos.

Ryan Brady
Ryan Brady
Founder, Digital Braid
|
·19 min read

A head of ops at a mid-sized B2B company sent me a Loom video in February. Thirty-eight seconds of one of her senior employees clicking through five tools to process a single customer request — CRM, spreadsheet, internal admin, email, back to the CRM to log the response. "We have eleven people doing this, forty hours a week each, and I just realized I can't remember the last time any of them actually spoke to a customer. They're data janitors now."

She had a bigger problem than the workflow. She had an organizational drift problem — eleven expensive hires slowly becoming software users instead of operators. The build we scoped wasn't about saving hours (though it saved plenty). It was about giving those eleven people their jobs back.

This is what custom AI tools actually do when they're built right. Not "replace humans with AI." Not "10x productivity." Put capable people back in front of the work only they can do, and let software handle everything else.

60–80%
Typical reclaim
Of hours a custom AI tool absorbs on a well-scoped operational workflow
4–12 wk
Focused build timeline
For a production system on a single workflow — not demos, actual shipping
$8K–$150K
Typical engagement range
Depending on scope: focused build, multi-workflow, or enterprise platform

This guide is for teams considering whether to build something custom — and for teams who've decided to but aren't sure what good looks like. It's written out of a decade of production engagements, including the ones that shipped late, the ones that didn't work the first time, and the ones we'd build differently today.


1. What Actually Counts as a "Custom AI Tool"

A custom AI tool is production software built around a language model for a specific workflow in a specific business. Not a chatbot. Not a Zapier flow with OpenAI pasted in. Not a Claude Project your team shares (though that's part of it). A real custom AI tool has three properties:

  1. It executes work — not just suggests or drafts. The output lands somewhere downstream where it matters (a CRM record, a customer-facing response, a published artifact, a transaction).
  2. It handles edge cases — because production data isn't the demo data. Real customers enter weird formats. Real workflows have exceptions. The system routes those to humans with enough context to resolve them quickly.
  3. It has observability — logs, evals, quality monitoring. Without it, silent regressions (a model update, a prompt drift, a data change) will tank output quality for weeks before anyone notices.

What a custom AI tool is NOT:

  • A demo that impresses leadership but nobody deploys
  • A SaaS product with AI features you configure slightly
  • A ChatGPT prompt your team copies into a tool
  • An RPA bot with "AI" bolted on the front

Those all have their place. None of them are what this guide is about.


2. Picking the Right Workflow — the Most Important Decision You'll Make

Teams don't fail because they build badly. They fail because they build the wrong thing. Workflow selection is the highest-leverage decision in the entire project, and most teams make it in about thirty minutes based on vibes.

The three questions that matter:

1. Does the workflow happen at least weekly, with clear inputs and outputs? Rare or bespoke workflows don't pay back automation overhead. Fuzzy inputs or fuzzy outputs make the build flaky. If a human can't describe the current workflow end-to-end in ten minutes, a machine can't run it yet.

2. Can you measure the current cost — hours, errors, revenue impact — in real numbers? Without a baseline you cannot prove ROI, size the build correctly, or know whether the finished system is working. The measurement itself is a gut-check: if your team can't tell you what the current workflow costs, they don't understand it well enough to automate it.

3. Is the workflow mostly pattern-matching, or mostly judgment? Pattern-matching work is LLM's sweet spot. Judgment work keeps humans in the loop — the LLM assists, but doesn't decide. Most real workflows are a mix, and the design question is where the handoff goes.


3. Build vs. Buy vs. Hybrid

This decision sits right next to workflow selection in importance. Most teams frame it as binary — we build or we buy. The correct framing is almost always hybrid: buy the commodity layers, build where the uniqueness is.

Buy when:

  • The problem is standard across the industry
  • You need something running this month
  • The SaaS tool's data model fits yours
  • Your competitive edge isn't in this workflow

Build when:

  • The workflow is specific to how you operate
  • Your volume makes SaaS per-seat pricing painful
  • Your data is sensitive or your stack has compliance constraints
  • You have (or can hire) someone to own the result

Hybrid when:

  • 80% of the workflow is standard, 20% is yours
  • You want to ship fast but own the critical path
  • You're in an industry where off-the-shelf is close but not close enough

The Build vs Buy tool scores eight factors across four axes and hands back a recommendation with an explicit "when this is wrong" caveat. Run it before scoping a real build — the thirty minutes it takes to fill in is often the thirty minutes that saves three months of wrong direction.


4. The Six Decisions That Shape the Build

Once you've picked the workflow and decided to build, the architecture falls out of six decisions. Get them right up front and the build is execution. Get them wrong and you'll rebuild.

1. Problem shape. Classification, extraction, generation, reasoning, or multi-step agent? Each shape calls for different patterns. Most teams default to "agent" because agents are new and impressive — but most real workflows are better served by simpler patterns.

2. Model choice. Claude, GPT, Gemini, open-source. Our default in 2026 is Claude Sonnet for production work — long context, careful reasoning, strong refusals — but the right answer depends on cost envelope, latency requirements, and compliance constraints. Don't let platform allegiance make this decision.

3. Context strategy. How does the model know what it needs to know? Options range from "prompt with examples" (fast, limited) to "RAG pipeline with vector search" (powerful, complex). Most teams over-reach for vector databases when a well-structured Claude Project with uploaded docs would do the job.

4. Interface. Where does a human meet this system? Slack bot, web app, API endpoint, email-based, inside an existing tool? The right interface is the one that's already in your team's workflow. Don't build a new surface if the work happens in one you already have.

5. Evaluation. How do you know the output is good? Evals — structured test suites — are what separate production systems from demos. Most teams under-invest here. Every model update and prompt change should run against evals before shipping.

6. Operations. Who runs this after it ships? Every engagement that produces a system with no post-launch owner watches that system rot within six months. Assign an owner before the build starts.


5. Where Custom AI Projects Actually Die

After watching dozens of builds across multiple companies, failure patterns cluster into a small number of buckets. Worth naming explicitly so you can avoid them.

Where roughly 80% of failures concentrate:

  • POC → MVP (scope + handoff): the demo ran on clean data in a sandbox. Production data is messy and lives in three systems that don't talk to each other. The team that built the demo hands off to an engineering team that doesn't own the original design. Work rots in transition.
  • MVP → Production (reliability + ownership): the system works for three users. It breaks at scale. Nobody owns it. Adoption stalls out.
  • Prompt/model regressions post-launch: a model update changes behavior silently. Evals would have caught it. There are no evals.
  • Business requirements changed: leadership pivoted. The workflow we automated is no longer the workflow we do. Happens.
  • Model or vendor killed the approach: rare, real. Anthropic deprecated Claude 3.5 in a week last month. We had to swap.

The POC→MVP gap is the deadliest because it looks like a technical problem when it's actually a scoping and ownership problem. The demo scoped against the interesting part of the work. Production requires scoping against the boring 80%.


6. Which AI Platform for Which Job

Before the stack, the platform choice. Most teams default to whichever tool they've been using, which is fine for personal productivity and wrong for production. Each platform has real strengths and real trade-offs that matter when you're building something clients or employees will depend on.

Claude (Anthropic)

Our default for production builds. Strong reasoning, long context windows (200K+ tokens on Sonnet, 1M on the experimental tier), careful refusals on ambiguous cases, and — importantly — the cleanest developer experience for agent work.

Key Claude products to know:

  • Claude Sonnet 4.5 — the production workhorse. What we default to for real builds. Same cost as earlier Sonnets ($3/M input, $15/M output), meaningfully better at following complex instructions.
  • Claude Projects — Claude.ai's workspace feature. Attach knowledge files (hundreds of them, millions of words combined) plus custom instructions and Claude persists that context across every conversation in the Project. Our go-to for "set up an AI workspace for a team" engagements.
  • Claude Artifacts — Claude's interactive canvas. When you ask Claude to build something (a dashboard mockup, a React component, an SVG diagram, a document), it renders the output in a live pane you can iterate on. Underrated for rapid prototyping and design exploration — a marketer can mock up a landing page variant in minutes.
  • Claude Code — Anthropic's CLI + IDE agent for coding work. Runs locally, operates on your actual codebase, commits changes. The current state of the art for AI-assisted engineering; we use it daily.
  • Computer Use / Claude as agent — Claude's ability to operate a computer via screenshots and keystrokes. Still maturing; worth watching but not yet a reliable production primitive for most use cases.

Use Claude when: you need long context (legal docs, large codebases, research corpora), production reliability, careful reasoning on ambiguous inputs, or strong coding work. Also: when you want a clean API with predictable behavior across versions.

ChatGPT / OpenAI

The most widely-adopted platform, best ecosystem for consumer-facing tools. OpenAI ships new capabilities faster than anyone, which is both a strength (cutting-edge features) and a risk (deprecations, product shuffles, API churn).

Key OpenAI products to know:

  • GPT-4 / GPT-5 — the flagship API models. Competitive with Claude on most benchmarks; slightly different strengths. Better out-of-the-box multimodal handling (images, audio); slightly noisier on complex reasoning.
  • Custom GPTs — ChatGPT's equivalent to Claude Projects. Builder UI, knowledge files, Actions (API integrations), available to any ChatGPT Plus/Team/Enterprise user. Best for shipping simple internal assistants fast, without a developer involved.
  • ChatGPT Canvas — the side-by-side editing UI for long-form work. Similar intent to Claude Artifacts, implemented differently. Useful for content and code drafting when you want structured editing instead of chat.
  • Codex (reborn) — OpenAI's coding-agent CLI, revived in 2025. Competitive with Claude Code on most tasks; better at some languages, weaker on others. Worth trying both on your codebase to see which matches your stack.
  • Operator — OpenAI's browser-operating agent. Can book flights, fill out forms, run workflows inside web apps. Still early; better for personal automation than production.
  • The Assistants API — OpenAI's framework for building stateful AI agents with tools and persistent threads. Powerful but heavier than the chat-completions API; most production work we do stays on chat-completions.

Use ChatGPT/OpenAI when: you want the broadest ecosystem, fast access to new features, strong multimodal, or a team already on ChatGPT Enterprise. Also: when the shipping deadline is tight and Custom GPTs can solve 80% of the problem without code.

Gemini (Google)

Best long-context model on the market and deeply integrated with Google Workspace. Gemini is often overlooked because Google's go-to-market has been weaker than Anthropic's and OpenAI's, but the underlying models are genuinely competitive — especially on tasks involving enormous documents or tight Google-ecosystem integration.

Key Gemini products to know:

  • Gemini 2.5 Pro — Google's flagship. Up to 2M token context (by far the largest in production). Underrated for tasks where "just put the whole codebase / document set / transcript archive in the prompt" is the simplest solution.
  • Google AI Studio — the developer playground. Free tier is generous, lets you test prompts interactively before calling the API. Best-in-class for rapid prompt iteration during design. We use AI Studio as a scratchpad even for projects we ship on Claude or OpenAI.
  • Gemini in Workspace — embedded inside Docs, Sheets, Gmail, Meet, and Drive. If your team lives in Google Workspace already, the per-seat Gemini upgrade turns every document into an AI-capable surface. Different value proposition than Claude/ChatGPT — it's ambient, not a separate destination.
  • Gems — Google's equivalent to Custom GPTs / Claude Projects. Available to Gemini Advanced subscribers. Weaker tooling UI than Custom GPTs, but integrates with Drive content natively.
  • Vertex AI — Google Cloud's enterprise platform for deploying Gemini (and other models) with governance, compliance, and scale controls. If you're on GCP already, this is the production path.

Use Gemini when: you need massive context windows, your team is deep in Google Workspace, you want the best free-tier developer playground (AI Studio), or you need enterprise compliance via Vertex AI.

Comparison at a Glance

Production agents + long docs
Claude Sonnet 4.5 — careful reasoning, strong tool use, good API ergonomics
Widest ecosystem + fastest shipping
ChatGPT / OpenAI — Custom GPTs, largest community, most third-party integrations
Long-context king + Google-native
Gemini 2.5 Pro — 2M token context, deep Workspace integration, AI Studio for dev iteration
Coding agents
Claude Code or OpenAI Codex — roughly tied; try both on your actual codebase
Interactive design / prototyping
Claude Artifacts — render + iterate on UI, code, diagrams, and documents in a live pane
Personal productivity automation
ChatGPT Operator — browser-operating agent with the most maturity in market
Free developer playground
Google AI Studio — generous free tier, fast iteration loop, no credit card to start
Embedded in existing tools
Gemini in Workspace or Copilot in Microsoft 365 — depending on which suite you live in

7. The Rest of the Stack

Platform picked — the surrounding infrastructure looks like this in 2026:

Integration layer: MCP (Model Context Protocol). Anthropic's open standard for connecting LLMs to tools and data sources. Replaces per-integration glue code with a uniform protocol. Still early, but standardizing fast — OpenAI and Google now both ship MCP support. When a new system needs to plug in, MCP shortens the work by weeks.

Orchestration layer: custom TypeScript + Python. Multi-step workflows, retries, error handling, human-in-the-loop handoffs. We avoid heavy orchestration frameworks (LangChain, LlamaIndex) for most production work — the abstraction tax rarely pays off. Plain code with good observability wins.

Context layer: Postgres with pgvector, or Pinecone for scale. We default to Postgres + pgvector for most engagements. It's a boring database your team can already operate. Pinecone and other managed vector DBs enter the picture when scale or latency requirements actually demand them — which is less often than teams assume.

Observability layer: LangFuse or Helicone. Prompt logging, trace capture, eval runs. This is the layer that separates a system that works today from a system that still works in six months. Do not skip.

Deployment: Vercel for UI, AWS Lambda / Cloud Run for agents. Everyday tooling. Nothing fancy. The model is the novel part; everything around it is boring infrastructure, and that's the point.


8. A Real Build, Compressed

The anonymized version of a project we shipped in late 2025:

The workflow: commercial door installer processing ~400 custom orders per month. Each order has 40+ specification fields — width, height, fire rating, hardware prep, glass type, frame depth, core material. A single wrong digit means a door gets manufactured to the wrong spec, ships across the country, and gets scrapped. Six-figure annual rework line item.

The old process: estimators keyed orders across three disconnected systems, cross-referenced spec sheets manually, caught most errors but not consistently. Senior estimators were the QA layer. New hires missed edge cases.

The build: an AI validation layer between the order form and the manufacturer's ordering system. Claude Sonnet cross-checks every specification against the live product database, validates dimensional tolerances, flags incompatible hardware configurations, and compares the order against the project's master spec file. Anything ambiguous routes to an estimator with a specific flag and suggested fix — not a generic warning.

The first version was embarrassing. Our initial validation logic caught 70% of errors but also false-flagged about 20% of valid orders. Estimators hated it and ignored the flags within two weeks. The fix was a better context layer — we trained on three months of the client's historical orders, not just the product database, so the system understood the patterns specific to their business.

Six months in: 97% error reduction, $400K+ annual savings, 80% faster order processing. The estimators got their Friday afternoons back. The factory stopped scrapping doors.

We used to run the QA check ourselves. Now the system runs it and flags the fifteen orders a month that actually need us. I'm finally doing the work I was hired for again.

Senior Estimator · Commercial Door Installer · Anonymized

The lesson that transferred: the win wasn't the model. The win was the context layer — three months of the client's real orders as training context, plus the product database, plus the master project files. A generic validator would have matched the demo performance. The client-specific context is what made it a production system.


9. What to Do Monday Morning

If you're reading this considering a custom AI build, here's the practical next move — not a call, an experiment.

  1. Pick a workflow. One. The most painful, most repetitive, most measurable one. Not "let's automate everything."
  2. Measure the baseline. Hours per week. Error rate. Cost per mistake. Team members involved. Write it down.
  3. Run a Friday audit. Spend two hours shadowing the team that does it. Document every handoff, every exception, every decision point. This is the boring step. This is also the step that separates projects that ship from projects that die.
  4. Score it with the tools. ROI Calculator for the hours-and-dollars math. Build vs Buy to decide direction. AI Readiness Assessment to check whether your team can execute.
  5. Scope conservatively. If your first scope is "automate the whole department," re-scope to the smallest valuable slice. The win compounds from there.

Have a workflow you think is a good candidate? Run the ROI Calculator for the math, check your readiness, or book a discovery call and we'll scope it together.

Share:

Get our insights

Perspectives on AI implementation, automation strategy, and marketing engineering. Delivered when we have something worth saying.