Skip to content
20 changes: 19 additions & 1 deletion docker-compose.components.yml
Original file line number Diff line number Diff line change
Expand Up @@ -114,7 +114,6 @@ services:
# Optionally, limit the GPUs exposed to the server.
# At least one GPU is required to run the YOLO TensorRT engine.
- NVIDIA_VISIBLE_DEVICES=all

command: [tritonserver,
--model-repository=/models,
--strict-model-config=false,
Expand Down Expand Up @@ -221,6 +220,25 @@ services:
image: ${REGISTRY}openmpf_ortools_subject_component:${TAG}
build: ${OPENMPF_PROJECTS_PATH}/openmpf-components/python/OrToolsSubjectComponent

llm-speech-summarization:
<<: *component-base
image: ${REGISTRY}openmpf_llm_speech_summarization:${TAG}
build: ${OPENMPF_PROJECTS_PATH}/openmpf-components/python/LlmSpeechSummarization

llm-speech-summarization-server:
image: ${REGISTRY}openmpf_llm_speech_summarization_server:${TAG}
build:
context: ${OPENMPF_PROJECTS_PATH}/openmpf-components/python/LlmSpeechSummarization
dockerfile: Dockerfile.vllm
deploy:
mode: global
# resources:
# reservations:
# devices:
# - driver: nvidia
# device_ids: ['0']
# capabilities: [gpu]

scene-change-detection:
<<: *component-base
image: ${REGISTRY}openmpf_scene_change_detection:${TAG}
Expand Down