You signed in with another tab or window. Reload to refresh your session.You signed out in another tab or window. Reload to refresh your session.You switched accounts on another tab or window. Reload to refresh your session.Dismiss alert
Why is this evaluation interesting?
This focuses on 16 African languages, evaluated on three knowledge QA and reasoning tasks such as AfriMMLU, AfriMGSM and AfriXNLI, human translated from MMLU, MGSM and XNLI respectively.
This sounds like a great idea, do you want to add them yourself to the library and open a PR? We've got a guide here for adding tasks and one here specific to multilingual evaluations :)
See: the tasks.py
Currently only the swahili subset is available, because we don't have correct translations for anchor words, that are required during the evaluation (e.g. answer/question).
If you are a native speaker in any of the languages we would love your help in improving that!
See the guide how to add a translations:
Evaluation short description
Why is this evaluation interesting?
This focuses on 16 African languages, evaluated on three knowledge QA and reasoning tasks such as AfriMMLU, AfriMGSM and AfriXNLI, human translated from MMLU, MGSM and XNLI respectively.
How used is it in the community?
Evaluation metadata: IrokoBench
Provide all available
Evaluation metadata: Uhura
Provide all available
Evaluation metadata: SIB-200
Provide all available
@NathanHB
The text was updated successfully, but these errors were encountered: