Overview

Details of specific measurements used to assess and quantify the resource utilization of LILT applications. These metrics provide insights into the performance, capacity, and scalability requirements of the applications. They include parameters such as CPU utilization, memory consumption, disk size and concurrent user sessions. By establishing baseline measurements and setting thresholds, usage-sizing metrics help identify normal operating levels and define boundaries for resource usage. These metrics enable proactive capacity planning, resource optimization, and the ability to detect anomalies or performance bottlenecks. The table below is meant to be a reference guide, and not an exact match to your individual environment, as each installation may choose to enable or disable specific services.

Usage-Sizing Metrics

App NameMin Recommended vCPUsMax Recommended vCPUsMin Recommended Memory (GB)Max Recommended Memory (GB)
Front0.5248
Beehive0.1444
Converter1122628
QA0.5134
Search0.05n/a2122
TM4123032
TB1n/a3030
Indexer0.05n/a2122
Lexicon483030
Watchdog0.1n/a23
Segment0.5244
File-translation361415
Job0.5123.5
Tag0.5123.5
Linguist0.5112
Auditlog0.5123.5
Assignment1244
Workflow1222
File-job241616
memory1244.6
notification0.5n/a24
auth0.120.1281
core-api0.120.1281
configuration-api0.120.1281
events-consumer0.120.1281
login0.120.1281
manager-ui0.120.1281
plugin-api0.120.1281
token-proxy0.120.1281
webhooks0.120.1281
webhooks-consumer0.120.1281
monitor0.120.1281
updatev3148080
Langid0.511.51.5
update-managerv30.250.511
Routing0.531.51.5
batchv30.250.511
batch-tb272230
llm-inference115430
Localpv0.120.1281
nginx-ingress0.10.128n/an/a
isitiod0.50.522
istio-ingressgateway0.120.1281
kiali0.010.010.0641
redis0.20.6410
rabbitmq1244
mysql0.5n/a48
mongoDB0.50.50.50.5
minIO0.10.511.5
clickhouse0.5n/a116
elasticsearch***1244
batch-worker-gpuv3 ****121212
translatev3 ****122539
Total*50 vCPUs160 vCPUs630 GB RAM728 GB RAM
Total*1 TB Disk****8 GPU**
*The following specifications are recommended for installing LILT on bare-metal hardware to support the given applications. It is important to note that these specifications may need to be adjusted based on the specific load requirements.
**To efficiently handle parallel batch requests, it is recommended to have the same number of GPUs as the number of required concurrent batch requests. This enables the system to effectively distribute the workload and process multiple requests simultaneously, leveraging the processing power of each GPU to improve throughput.
*** These values are doubled, because there are two replicas
**** These pods run on the GPU node
n/a marked services have no maximum limit and can use the available resources on the node