Skip to main content

Deployment Overview

This page walks through the high-level deployment process and helps you choose the right deployment model for your infrastructure.

Deployment Steps

Regardless of the deployment model, every installation follows these steps:

1

Get the container image

Authenticate with the registry and pull the image. See the Container Image guide.

2

Choose a deployment model

Select the platform that matches your infrastructure. See the options below.

3

Configure and deploy

Follow the platform-specific deployment guide to run the engine.

4

Verify the service

Confirm the engine is healthy and responding to gRPC requests. See Monitoring & Logging.

5

Secure the deployment

Apply network policies and security best practices. See Security.

Choosing a Deployment Model

Kubernetes-based

Best for production workloads that need horizontal scaling, rolling updates, and self-healing. The Hiya engine is stateless and scales by simply increasing the replica count.

Managed Container Services

Suitable when you want the cloud provider to manage the underlying compute or when Kubernetes is not part of your stack.

Single Host

Best for development, testing, or low-traffic deployments on a single machine.

Common Configuration

All deployment models share these configuration parameters:

Environment Variables

VariableRequiredDefaultDescription
API_KEYYesProvided by Hiya. Used for usage metering.
ORG_HANDLEYesYour organization handle. See List your organizations.
PORTNo8080gRPC API port.
WS_PORTNo8081WebSocket API port.

Model Storage

At startup, the engine loads ML models into the path /opt/loccus/models. This path should be backed by a RAM-based filesystem (tmpfs, emptyDir with medium: Memory, or equivalent) so that model data is held in memory and never written to persistent disk.

Network Requirements

DirectionDestinationPortProtocolPurpose
Outboundapi.hiya.com443HTTPSLicense verification and billing
InboundEngine container8080gRPCVoice verification API
InboundEngine container8081WebSocketWebSocket API

Using the Service

The self-hosted engine exposes the same API interface as the Hiya cloud service. This means you can:

  1. Develop and test your integration against the cloud API at wss://api.hiya.com/audiointel/...
  2. Deploy the self-hosted engine using one of the guides above
  3. Switch your client to point at the self-hosted endpoint by changing the host and port

The engine supports the same WebSocket streaming endpoint. The only difference is the destination — instead of api.hiya.com, you point to your self-hosted engine's address (e.g., ws://your-engine-host:8081).

We recommend developing your integration against the cloud API first, then switching to self-hosted once your deployment is verified. This lets you validate your client code independently of infrastructure setup.