You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
When evaluating large models, significant costs and delays can occur for inference, especially on larger datasets. Possibly I want to re-evaluate my predictions using different metrics.
Solution/Feature
I want the predictions to be saved in an inspectable cache which can be used when the evaluation is run again.
The text was updated successfully, but these errors were encountered:
If you use the different saving parameters (as indicated in the doc), your predictions (results and/or details) are saved and can be used for reinspection later on. The quickest way to get what you need is therefore using the details file to recompute the metrics on them by hand.
Since not all metrics use the same generation methods, we have not prioritized a cache atm (to prevent risks such as running a greedy eval, then a sampling one, and accidentally using the same results for metric computations), but we'll add your suggestion to our todo!
Issue encountered
When evaluating large models, significant costs and delays can occur for inference, especially on larger datasets. Possibly I want to re-evaluate my predictions using different metrics.
Solution/Feature
I want the predictions to be saved in an inspectable cache which can be used when the evaluation is run again.
The text was updated successfully, but these errors were encountered: