Skip to content
This repository was archived by the owner on Jul 24, 2025. It is now read-only.
Merged
Changes from all commits
Commits
File filter

Filter by extension

Filter by extension

Conversations
Failed to load comments.
Loading
Jump to
Jump to file
Failed to load files.
Loading
Diff view
Diff view
7 changes: 7 additions & 0 deletions README.md
Original file line number Diff line number Diff line change
@@ -1,5 +1,11 @@
# ModelService

> ❗ `llm-d-model-service` is a Kubernetes operator that manages ModelService objects in `llm-d-deployer`. This project is now deprecated and no longer under active maintenance. The [ModelService helm chart proposal](https://github.com/llm-d/llm-d/blob/dev/docs/proposals/modelservice.md) was accepted by llm-d on June 10, 2025. In summary, the core functionality previously handled by llm-d-model-service is now provided by a Helm chart.

> For the latest updates and continued development, please refer to:

👉 [llm-d-incubation/llm-d-modelservice](https://github.com/llm-d-incubation/llm-d-modelservice)

> *ModelService* declaratively provisions and maintains the Kubernetes resources needed to serve a base model for inference.

A *ModelService* custom resource encapsulates the desired state of workloads and routing associated with a single base model. It automates the management of Kubernetes resources, including:
Expand Down Expand Up @@ -91,3 +97,4 @@ The result is a fully managed inference stack for the base model.
7. Prometheus metrics exporter: Emit controller metrics

8. Enable multi-node inferencing: for instance, using LWS integration
# llm-d-modelservice
Loading