You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
See in the UI the traces that are produced when using LLM as Judges, both General and LLM spans.
Motivation
When using LLM as Judged metrics in your Evaluations, it is useful to track the Evaluation because:
Sometimes you want to iterate on the creation of the actual metric, and you need to analyze / compare the prompts that were used to evaluate the LLM-App.
It allows tracking of the full usage/cost of the Evaluation
The text was updated successfully, but these errors were encountered:
@jverre I like this idea. However, we need to decide how enable/disable it. We'll likely need to have some flags to disable/enable opik tracking for score functions
Proposal summary
See in the UI the traces that are produced when using LLM as Judges, both General and LLM spans.
Motivation
When using LLM as Judged metrics in your Evaluations, it is useful to track the Evaluation because:
The text was updated successfully, but these errors were encountered: