LLM Observability: What to Log, Monitor, and Alert On
A production LLM stack should log prompts, responses, model/version metadata, latency, token usage, refusals, and safety events so teams can detect drift, prompt injection, and cost spikes before users do. This post compares where Langfuse, Helicone, and Arize fit in the pipeline—and which signals each one surfaces best for alerting and anomaly detection.