Gist: LaunchDarkly adds online evaluations to AI Configs, using LLM-as-a-Judge to score completions in production. The feature measures accuracy, relevancy, and toxicity in real time and can trigger fallbacks when quality drops.
Signal reason: The post announces online evals and LLM-as-a-Judge as a new capability in AI Configs.
