On-Prem Features and Updates
Multimodal OCR Service
LILT 5.3 introduces a powerful new neural-based OCR service that delivers significantly improved accuracy for image and scanned PDF translation. The multimodal OCR service leverages advanced machine learning models to better extract text from complex documents, including those with mixed content types, tables, and challenging layouts. Key Benefits:- Enhanced accuracy for scanned PDFs and images
- Improved handling of complex document layouts
- Better recognition of tables and formatted content
- Seamless integration with existing translation workflows
values.yaml:
Enhanced LLM Inference Architecture
LILT 5.3 restructures the LLM inference system into dedicated, optimized services for different AI models. This architectural improvement provides better resource management, improved performance, and more granular control over AI capabilities.New Specialized Inference Services:
Llama vLLM Inference- Optimized inference for Llama language models
- GPU-accelerated processing (requires 4 GPUs)
- High-performance vLLM backend
- Resource requirements: 100G memory, 5-14 CPUs
- Real-time speech-to-text transcription
- GPU-accelerated for faster processing
- Supports multiple languages
- Ideal for audio/video translation workflows
- Dedicated inference for Gemma3 models
- Optimized for specific translation tasks
- GPU-accelerated processing
- Specialized inference for Emma-500 models
- Advanced language understanding capabilities
capability: gpu.
Third-Party MT Batch Processing
A new CPU-based batch worker specifically designed for third-party machine translation (3pMT) processing has been added. This dedicated worker improves throughput and reliability for integrations with external MT providers. Features:- Dedicated CPU-based processing for 3pMT
- Configurable worker pools
- 24G memory for large document handling
- Improved batch processing reliability
Enhanced Document Conversion
Document conversion timeouts have been made configurable, allowing better handling of large and complex documents. The new timeout settings apply to both the converter and file-translation services. Configuration:MySQL SSL Support
LILT 5.3 adds comprehensive SSL/TLS support for MySQL connections, enhancing security for database communications. SSL is now enabled by default with flexible configuration options. Configuration:REQUIRED: Forces SSL connections (recommended for production)PREFERRED: Uses SSL if available, falls back to unencryptedDISABLED: Disables SSL (not recommended)
Core Platform Updates
Internal Architecture Improvements
Service Naming Standardization- AV scanner service renamed to
lilt-av-scanfor consistency - ConfigMap and Secret resources standardized with
lilt-prefix - Improved service discovery and management
Breaking Changes
LLM Inference Service Restructuring
The genericllm-inference service has been removed and replaced with specialized inference services. This change improves resource utilization and provides better control over AI model deployment.
Migration Required:
If you previously had llm-inference enabled, you must:
-
Disable the old service:
-
Enable the appropriate replacement services:
-
Update GPU node labels if not already configured:
ConfigMap and Secret Name Changes
Several services now reference standardizedlilt-configmap and lilt-secrets instead of service-specific resources.
Affected Services:
- Translation Memory (tm)
- Lexicon
-
Ensure your deployment creates resources with the new names:
lilt-configmap(previouslyfront-configmap)lilt-secrets(previouslyfront-secrets)
-
If using custom Helm templates, update references:
Init Container File Path Changes
Services using the two-stage init container pattern (lexicon, file-job) must update file path configurations. Old Format:s3://lilt/ prefix.
New Configuration Requirements
Required Secrets
Core-API S3 Credentials
Add tosecrets.yaml:
Multimodal OCR (If Enabling)
Add tosecrets.yaml:
Optional Configuration Updates
Backend Configuration Toggle
A new toggle allows enabling/disabling backend configuration features:BigQuery Integration
For customers using BigQuery analytics:Batch Worker Memory
Batch worker memory limits have been increased for improved performance:Infrastructure Requirements
GPU Node Requirements
If enabling LLM inference services, ensure GPU nodes are available:| Service | GPU Count | Memory | CPU |
|---|---|---|---|
| llama-vllm-inference | 4 | 100G | 5-14 |
| llm-inference-whisper | Yes | Variable | Variable |
| gemma-vllm-inference | Yes | Variable | Variable |
| emma-500-vllm-inference | Yes | Variable | Variable |
capability: gpu
Setup:
Worker Node Requirements
CPU-intensive services require worker nodes:| Service | Memory | CPU |
|---|---|---|
| batch-worker-cpu-3pmt | 24G | 1 |
| Other batch workers | Variable | Variable |
node-type: worker
Setup:
Troubleshooting
LLM Inference Services Not Starting
Symptom: LLM inference pods stuck inPending state
Solution:
-
Verify GPU nodes are available:
-
Check GPU resources:
-
Ensure proper node labels:
Multimodal OCR Connection Issues
Symptom: Multimodal service cannot connect to RabbitMQ Solution:-
Verify RabbitMQ credentials in
secrets.yaml: -
Check RabbitMQ connectivity:
Batch Worker Memory Issues
Symptom: Batch workers being OOMKilled Solution:-
Verify node capacity:
-
Increase memory if needed:
ConfigMap/Secret Not Found Errors
Symptom: Services failing with “configmap ‘lilt-configmap’ not found” Solution: Ensure the main Lilt chart creates these resources. Check your Helm templates include:S3/MinIO Connection Issues
Symptom: Services cannot access S3/MinIO storage Solution:-
Verify S3 credentials in
secrets.yaml: -
Verify internal MinIO endpoint is accessible:

