Resource requirements for deploying Onyx
Resource | Minimum | Preferred |
---|---|---|
CPU | 4 vCPU | 8+ vCPU |
RAM | 10 GB | 16+ GB |
Disk | 50 GB + ~2.5x indexed data | 500 GB for organizations <5000 users |
docker system prune --all
.
Provider | Recommended Instance Type |
---|---|
AWS | m7g.xlarge |
GCP | e2-standard-4 or e2-standard-8 |
Azure | D4s_v3 |
DigitalOcean | Meet the preferred resources in the table above |
Vespa on older CPUs
vespaengine/vespa-generic-intel-x86_64
image in your Docker Compose file.
This generic image is slower.For more details, see Vespa CPU Support.values.yaml
for our default requests and limits.
Component | CPU | Memory |
---|---|---|
api_server | 1 | 2 Gi |
background | 2 | 8 Gi |
indexing_model_server | 2 | 4 Gi |
inference_model_server | 2 | 4 Gi |
postgres | 2 | 2 Gi |
vespa | >= 4 | >= 8 Gi |
nginx | 250m (1/4) | 128 Mi |
vespa
recommendation is the bare minimum for a production deployment.
With 50GB of documents, we recommend at least 10 CPU, 20Gi Memory.index
component will need:
m7g.4xlarge
or a c5.9xlarge
EC2 instance would be appropriate.
If deploying with Kubernetes or AWS EKS, this would give a per-component resource allocation of:
Component | CPU | Memory |
---|---|---|
api_server | 1 | 2 Gi |
background | 2 | 8 Gi |
indexing_model_server | 2 | 4 Gi |
inference_model_server | 2 | 4 Gi |
postgres | 2 | 4 Gi |
vespa | 10 | 34 Gi |