Skip to content
Skip to content

Technology

The platform. Explained without the marketing layer.

WiseAI Agency is a vertical SaaS engine — one voice AI stack, one web stack, deployed across multiple industry verticals. Here is exactly how it works.

How it works

Six layers. Every vertical uses all of them.

Voice infrastructure — LiveKit Cloud

Every WiseAI vertical runs on LiveKit Agents, a production-grade voice AI framework. One agent pool handles all customers across all verticals simultaneously — multi-tenant by design. Each customer's session loads their specific configuration at call time.

We use Deepgram Nova-3 for speech-to-text, with per-vertical keyterms boosted at session start. Cartesia Sonic handles text-to-speech. The voice model does not change; the content, prompts, and routing rules do.

LiveKit Agents 1.5Deepgram Nova-3Cartesia SonicSIP/PSTN via Telnyx

Language models — not one, several

Coordinator and Care agents run Claude Haiku 4.5 with prompt caching — the model that handles the first moments of a call. Caching means the system prompt (thousands of tokens) is loaded once per customer session, not re-transmitted on every turn. Gemini 2.5 Flash handles fallback when primary is unavailable.

No model has access to prior customers' data. Each session is isolated. System prompts are assembled at call start from per-vertical templates + per-customer config + product knowledge loaded from the database.

Claude Haiku 4.5 (prompt caching)Gemini 2.5 Flash (fallback)Per-session isolation

Multi-agent orchestration

A Coordinator agent handles intake and routing. A Care agent handles emotionally sensitive conversations. A Sales agent handles product inquiries and lead capture. Each role has its own system prompt, its own tool access, and its own escalation rules.

The Coordinator decides which agent speaks next based on call context. In a funeral home context, the Care agent is primary for 90% of calls. In a church context, the Coordinator handles most general inquiries, the Care agent handles grief and pastoral requests.

Coordinator AgentCare AgentSales AgentContext-driven handoff

HEAR protocol — the moat

Most voice AI platforms have a technology moat. Ours is not just technical — it is pastoral. The HEAR protocol (Hear, Empathize, Advance, Respond) came from Clinical Pastoral Education and Stephen Leader training. It encodes how trained care professionals listen to people in distress.

The protocol is baked into the Care agent's system prompt at the prompt-fragment level — not bolted on as an instruction. Every care-sensitive interaction follows it by default. This is the product behavior that competing platforms cannot copy without lived experience.

HEAR Protocol (Clinical Pastoral Education)Care agent system promptCrisis detection layer

Per-vertical training

Each vertical has its own product knowledge table in the database. Funeral homes load a different knowledge base than churches. At call start, the agent's system prompt is assembled from: the vertical's base template + the customer's specific configuration (services, pricing, staff contacts, chapel names) + the most relevant product knowledge rows.

Knowledge is updateable by the customer through a dashboard. Changes take effect on the next call. No redeployment required.

Supabase + vector search (RAG)Per-customer product knowledgeReal-time config at session start

Chatbot + web layer

Alongside voice, every vertical offers a web chatbot built on the Vercel AI SDK 6 with streaming SSE. The same HEAR protocol applies. The same knowledge base. Tool access is gated by plan tier — premium customers can unlock FAQ lookup, appointment capture, and escalation triggers.

The web layer is a Next.js 16 App Router application deployed on Vercel. ISR-rendered public pages load in under 200ms globally. The admin dashboard gives each customer real-time access to call logs, transcripts, and performance metrics.

Next.js 16 App RouterVercel AI SDK 6Supabase + StripeVercel Edge Network

Engineering decisions

The decisions that define the platform.

One codebase. Not a fork per customer.

Multi-tenancy is load-bearing. Every customer shares the same deployed agent. Per-customer behavior is config, not code. This means updates ship to all customers simultaneously and bugs are fixed once.

Managed platforms, not custom infrastructure.

LiveKit manages the voice infrastructure. Vercel manages the web layer. Supabase manages the database. We build on the best managed platforms available — not because it is easier, but because it is the right engineering decision for a focused product team.

Privacy by default.

No customer's data is used to train any model. Calls are retained per the customer's configuration. Session isolation is enforced at the framework level. The service role key never reaches a browser.

The technology serves the human moment, not the other way around.

The reason we use Claude Haiku with prompt caching is response latency. A family calling a funeral home at 2 a.m. cannot wait two seconds for a response to load. Every technical decision traces back to the human experience at the end of the call.

Want to see the technology on your use case?

If you are in a vertical we have productized, book a demo. If you are in a vertical we have not, get in touch — we will tell you honestly whether the engine fits.