KServe Monitoring Card
The KServe Status card (kserve_status) provides real-time visibility into KServe model serving infrastructure across your managed clusters.
What It Shows
- InferenceService status — Ready/Not Ready state for each deployed model
- Replica counts — Current vs desired replicas per InferenceService
- Latency metrics — P50/P95 request latency where available via DCGM or Prometheus
- Error rates — Inference request failures per InferenceService
- Framework — Model framework (TensorFlow, PyTorch, ONNX, etc.)
Adding the Card
- Open Console Studio from the sidebar (the Add more… button)
- Search for KServe Status
- Drag it to your dashboard or click Add
Requirements
- KServe operator installed on managed clusters
ClusterRolewith read access toserving.kserve.ioresources- Optional: Prometheus/DCGM metrics for latency data
Demo Mode
When no live KServe installation is detected, the card displays demo data showing a representative set of InferenceServices with varying status states.