A sovereign voice AI appliance — 2U, on-prem. A single chassis ships the entire voice AI platform inside: real-time and recorded speech-to-text, language identification, and operational APIs — in your rack, under your control.
Why operators choose Auricus Voice
- Sovereignty & control — audio and transcripts stay inside your network. Designed for finance, healthcare, public sector, defense, and any deployment where cross-border processing or third-party AI providers are a non-starter.
- Predictable economics, end-to-end — single capital line item. No per-minute API meter, no egress, no rate-limit surprises. See the savings analysis for the matched-workload comparison.
- Reliability — no dependency on external API uptime or internet jitter. Latency is bounded by your rack.
- Integration — REST ingestion, async job queues, results by poll or webhook. Drops into existing ETL, CRM, and analytics stacks.
- Operational visibility — Prometheus metrics + Grafana dashboards, structured audit logs, exportable SLO tracking.
Product family
Same model, same pipeline, same APIs across the family. Pick chassis density to match workload.
| SKU |
Concurrent live calls |
Matched annual audio (M min/yr) |
Batch (files/min) |
End-to-end (files/hr) |
Peak accelerator power |
Target use case |
| Auricus Voice 8 |
12 |
2 |
25 |
60–75 |
80 W |
Mid-size contact center, departmental fleet |
| Auricus Voice 16 |
24 |
4 |
50 |
120–150 |
160 W |
Large enterprise contact center, regional carrier |
| Auricus Voice 32 |
48 |
8 |
100 |
240–300 |
320 W |
National contact center, telco / government scale |
Concurrent live calls = sustained concurrent near-real-time conversations per appliance. Matched annual audio = realistic operational capacity per appliance under a typical mixed real-time + batch duty cycle. Batch (files/min) = sustained transcription throughput per appliance. files/hr (E2E) = end-to-end including ingest, language ID, decode, and delivery overhead. Peak accelerator power scales linearly across the family; add ~120 W chassis baseline for total appliance draw.
→ Public spec sheet: Specifications · platform deep dive: Platform.
- Live and near–real-time — windowed streaming for contact centers, supervision, and compliance workflows. ~48–64 concurrent near–real-time streams per Auricus Voice 32 appliance before queue limits dominate.
- Batch at scale — sustained ~100 files/min on Auricus Voice 32; end-to-end (ingest, language ID, delivery) ~240–300 files/hour under typical overhead.
- Languages — ~99 transcription targets and 107 automatic language-ID classes.
- Auricus Voice 32 at full load: ~600 W typical appliance draw — vs comparable GPU-based racks measured in kilowatts.
- 2U chassis — one rack-unit-class enclosure replaces multi-server GPU racks, cutting embodied materials, cooling load, and end-of-life e-waste.
Compliance posture
- EU — AI Act 2 August 2026 ready — vendor-side primitives for Articles 9–17 (provider) and Article 26 (deployer) obligations under Regulation (EU) 2024/1689 where the deployer’s use case is high-risk; GDPR-aligned with on-prem retention, structured audit logs, and the architecture proposed in the EU Digital Omnibus.
- US — sector-ready — primitives for HIPAA (BAA available), GLBA Safeguards Rule, PCI DSS 4.0.1, and CJIS; mappings to NIST CSF 2.0, NIST AI RMF 1.0, NIST SP 800-53 Rev. 5 and 800-171 Rev. 3 / CMMC 2.0 for federal and defense customers.
- State-level — biometric-privacy aware (Illinois BIPA, Texas CUBI, Washington H.B. 1493), CCPA / CPRA sensitive-PI ready, and aligned with the Colorado AI Act (in force 1 Feb 2026).
- On-prem only — no third-party processor in the inference path; no external egress required.
- TLS in transit, bearer-token authentication on the REST API, structured audit logs with request IDs.
→ Full posture (EU + US): Compliance.
Get started
ask@auricus.com — pilots, pricing, and integration questions.