How to Monitor LLM Inference on Kubernetes with OpenTelemetry | DailyDevLists