02 โ€” Hybrid AI engine

Your AI. Your choice. Your data.

Five AI providers, hybrid orchestration, and BYOK. Pick the best model for each task โ€” or run entirely on-premise with Ollama for zero data egress.

5
AI providers
BYOK
Bring your own key
On-premise
Ollama support
Auto
Failover & routing
Zero
Data egress option
Supported providers

Five providers. One platform.

Provider Models Best for Data egress Cost
โšก Groq Llama-3.3-70b, Llama-3.1-8b/70b Speed-critical, high-volume Cloud Per token
๐Ÿค– OpenAI GPT-4o, GPT-4.1, GPT-4.1-mini General intelligence, complex reasoning Cloud Per token
โœจ Google Gemini Gemini-2.5-Flash/Pro, Gemini-1.5 Multimodal, long context windows Cloud Per token
๐Ÿง  Anthropic Claude Claude-3.5-Sonnet, Haiku, Opus Careful reasoning, safety-focused Cloud Per token
๐Ÿ  Ollama (local) On-premise Llama3, Mistral, Phi3, Gemma2 Sensitive data, zero API cost None Free
Ollama on-premise

Zero data egress. Zero API cost.

Run AI models entirely on your own servers. No data ever leaves your infrastructure. Ideal for healthcare, finance, legal, and any business with sensitive data requirements.

๐Ÿ”’
Complete data sovereignty

Patient records, financial data, legal documents โ€” none of it leaves your network.

๐Ÿ’ฐ
Zero per-token cost

Run unlimited queries on your own hardware. No API bills, no usage caps.

โšก
Full AI capability

Same quality responses as cloud AI โ€” Llama3, Mistral, Phi3, Gemma2 on your hardware.

๐Ÿ”ง
Per-task model assignment

Assign different local models for entity extraction, response generation, summarization, and quality scoring.

Supported Ollama models

Ll
Llama3
General purpose, excellent reasoning
Meta
Mi
Mistral
Fast, efficient, great for chat
Mistral AI
Ph
Phi3
Small but powerful, low resource usage
Microsoft
Ge
Gemma2
Balanced performance and speed
Google
Cu
Custom models
Any Ollama-compatible model you deploy
Any
Intelligent routing

The right model for every query

๐Ÿ”‘

BYOK

Bring your own API key for any provider. Tenant-level isolation โ€” your key, your usage, your billing. Never shared.

๐Ÿ”„

Primary + backup

Set a primary provider and a backup. Automatic failover if the primary goes down or rate-limits.

๐Ÿ“Š

Confidence routing

Queries are automatically routed to the best engine based on confidence scoring and query complexity.

โš–๏ธ

Load balancing

Distribute queries across multiple providers or models to optimise cost and performance.

Choose your AI. Deploy in minutes.

All 5 providers available on every plan. Ollama on-premise available on Pro and Enterprise.