-
Notifications
You must be signed in to change notification settings - Fork 0
Expand file tree
/
Copy pathenv.template
More file actions
82 lines (65 loc) · 2.74 KB
/
env.template
File metadata and controls
82 lines (65 loc) · 2.74 KB
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
41
42
43
44
45
46
47
48
49
50
51
52
53
54
55
56
57
58
59
60
61
62
63
64
65
66
67
68
69
70
71
72
73
74
75
76
77
78
79
80
81
82
# =============================================================================
# OpenProcessor Configuration Template
# =============================================================================
# Copy this file to .env: cp env.template .env
# Or run ./scripts/setup.sh to auto-generate based on your GPU
# =============================================================================
# =============================================================================
# GPU Configuration
# =============================================================================
# GPU Profile (auto-detected during setup)
# Options:
# minimal - 6-8GB VRAM (RTX 3060, RTX 4060) - core models only
# standard - 12-24GB VRAM (RTX 3080, RTX 4090) - all models
# full - 48GB+ VRAM (A6000, A100) - maximum parallelism
GPU_PROFILE=standard
# Which GPU to use (0-indexed)
# Use nvidia-smi to see available GPUs
GPU_ID=0
# =============================================================================
# Performance Settings (auto-configured based on GPU_PROFILE)
# =============================================================================
# Number of uvicorn workers for the FastAPI service
# Adjust based on CPU cores and memory
# Recommended: minimal=8, standard=16, full=32
TRITON_WORKERS=16
# Maximum batch size for inference
# Lower values use less VRAM but reduce throughput
# Recommended: minimal=16, standard=32, full=64
MAX_BATCH_SIZE=32
# Shared memory size for Docker containers
# Used for inter-process communication
# Recommended: minimal=4g, standard=6g, full=8g
SHM_SIZE=6g
# =============================================================================
# OpenSearch Vector Database
# =============================================================================
# JVM heap size for OpenSearch
# Affects search performance and memory usage
# Recommended: minimal=1g, standard=2g, full=4g
OPENSEARCH_HEAP=2g
# =============================================================================
# Port Configuration
# =============================================================================
# Change these if ports conflict with other services on your system
# Main API port (FastAPI service - all endpoints)
API_PORT=4603
# Triton Inference Server ports
TRITON_HTTP=4600
TRITON_GRPC=4601
TRITON_METRICS=4602
# Monitoring stack
PROMETHEUS_PORT=4604
GRAFANA_PORT=4605
LOKI_PORT=4606
# Vector database
OPENSEARCH_PORT=4607
OPENSEARCH_DASHBOARDS_PORT=4608
# =============================================================================
# Development Settings
# =============================================================================
# Enable debug mode (verbose logging)
# DEBUG=1
# Log level for uvicorn
# Options: critical, error, warning, info, debug, trace
# LOG_LEVEL=info