You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Currently, we use the variant of recall described in Huang 2021:
Our initial results have been very promising. The primary metric we are currently using is recall, the percentage of documents defined as relevant that we are successfully able to recommend. We adopt the measurement approach taken by Huang et. al (2021).
We select a random opinion in the federal corpus and remove it from our network (as if the opinion never existed).
We input all but one of the opinion’s neighbors into the recommendation software.
We measure whether the omitted neighbor was the top recommendation, in the top 5 recommendations, or in the top 20 recommendations.
This is alright, but leaves a lot to be desired with respect to a fuller understanding of our models' performance and its ability to surface useful cases. We've got some other ideas (to be documented at a later time) of what kinds of metrics might better serve us.
The text was updated successfully, but these errors were encountered:
Currently, we use the variant of recall described in Huang 2021:
This is alright, but leaves a lot to be desired with respect to a fuller understanding of our models' performance and its ability to surface useful cases. We've got some other ideas (to be documented at a later time) of what kinds of metrics might better serve us.
The text was updated successfully, but these errors were encountered: