-
Notifications
You must be signed in to change notification settings - Fork 60
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Evaluations – Remove conflicts in evaluation objective between configuration and prompt #420
Comments
The planTo do this, the schema must change. Now, each evaluation will have 2 polymorphic relations: There will currently be 2
And 3
The evaluation will expect results depending on the This allows for many more types of evaluations in the future, both llmAsJudge or any other type (like Human in the Loop), while maintaining the resultable types that we have now.
Development breakdownPart 1 —
|
What?
Right now, there is a config when creating an evaluation that sets the result (numeric between 1 and 5), but we don't pass this to the prompt, so the user can set a range between 9 and 20, and it would be the one that takes into account.
In summary, there are 2 sources of truth.
https://www.figma.com/design/ODioXiqX8aeDMonsh0HBui/Latitude-Cloud?node-id=2738-34189&t=C31y3Hbykh3pzF2x-4
The text was updated successfully, but these errors were encountered: