Who this comparison is for
AI Hub highlights
- Unified gateway for OpenAI + Anthropic + Gemini + Mistral + HF
- Conversation memory, search (FTS5), ratings, snippets
- Usage dashboard + monthly token quotas, Redis memo-cache
- Moderation policy (block/warn/allow) with CSV audit export
OpenAI Platform highlights
- Native access to OpenAI models & features
- Best-effort tooling inside one provider ecosystem
Capability matrix
Capability | AI Hub | OpenAI Platform | Notes |
---|---|---|---|
Multi-provider routing (one API) | Full | None | AI Hub routes across multiple vendors; OpenAI is single-provider |
Fallback on 5xx (OpenAI→Anthropic) | Full | None | Configurable secondary provider |
Conversation memory + search | Full | Partial | AI Hub ships chat storage & FTS; OpenAI leaves storage to you |
Moderation policy (pre/post, audits) | Full | Partial | OpenAI moderation API exists; AI Hub adds policy + CSV export |
Usage dashboard (30 days) & quotas | Full | Partial | Hub enforces hard monthly token limits |
Jobs queue + progress events | Full | Manual | Requires custom worker when built direct |
Embeddings (HF wrapper optional) | Full | Native | OpenAI has native embeddings; Hub also supports HF |
Admin key vault (AES-256-GCM) | Full | Manual | Hub includes workspace-scoped vault |
Redis memo-cache for deterministic calls | Full | Manual | DIY when building direct |
Web UI (Chat, Keys, Playground, Audits) | Full | Partial | Hub ships a single-instance UI |
Events (AI.response.*, AI.job.*) | Full | Manual | Hub publishes internal bus events |
- Comparisons focus on orchestration & governance layers, not model quality.
- Matrix tokens: full/partial/none/native/via_zapier/manual indicate support depth.
Total cost of ownership
Direct OpenAI integration is fast for a single feature. As usage scales, teams add storage, queues, quotas, and audits. AI Hub provides these out-of-the-box, reducing build/maintain costs and enabling provider choice.
Assumptions
- 3–5 AI features across products
- Need for quotas, audits, fallback, and team chat UI
Migration plan
From OpenAI Platform · Wrap calls via AI Hub Query API; move chat to Hub; enable quotas & audits
-
1
Store OpenAI key in AI Hub KeyVault; set primary provider
-
2
Switch endpoints to /QueryApi.php and validate outputs
-
3
Enable moderation policy + UsageDaily quotas
-
4
Optionally add Anthropic/Gemini as secondary fallback
Security
- Workspace-scoped key vault, AES-256-GCM at rest, TLS in transit
- SAML-gated audit CSV export
Evidence & sources
Claim | Value | Source |
---|---|---|
Unified gateway & quotas | Query API + UsageDaily + RateLimiter |
product_docs
Provider-agnostic layer
|
About AI Hub
AI Hub centralizes generative AI for your workspace with a single, policy-aware gateway to multiple providers. Teams get a streamlined chat experience with searchable history and feedback, a minimal Query API for quick prompts, and embeddings for retrieval workflows. Operators gain visibility with usage & cost tracking, quotas, and exportable audit logs.
Choose the best model for each task, fail over between providers, and moderate inputs/outputs with block/warn/allow policies. Keys are encrypted at rest and scoped per workspace. Long-running tasks run on a background worker and broadcast events so other apps can react in real time.
Designed for safety and speed: opinionated defaults, least-privilege access, and drop-in APIs that make it easy to bring AI to every surface of Velaxe.