Skip to content
New issue

Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.

By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.

Already on GitHub? Sign in to your account

[EVAL]: Add more African Benchmarks #373

Open
dadelani opened this issue Oct 24, 2024 · 4 comments
Open

[EVAL]: Add more African Benchmarks #373

dadelani opened this issue Oct 24, 2024 · 4 comments
Labels
good first issue Good for newcomers help wanted Extra attention is needed new task

Comments

@dadelani
Copy link

dadelani commented Oct 24, 2024

Evaluation short description

  • Why is this evaluation interesting?
    This focuses on 16 African languages, evaluated on three knowledge QA and reasoning tasks such as AfriMMLU, AfriMGSM and AfriXNLI, human translated from MMLU, MGSM and XNLI respectively.

  • How used is it in the community?

Evaluation metadata: IrokoBench

Provide all available

Evaluation metadata: Uhura

Provide all available

Evaluation metadata: SIB-200

Provide all available

@NathanHB

@clefourrier
Copy link
Member

Hi @dadelani , thanks for your issue!

This sounds like a great idea, do you want to add them yourself to the library and open a PR? We've got a guide here for adding tasks and one here specific to multilingual evaluations :)

@hynky1999
Copy link
Collaborator

hynky1999 commented Oct 24, 2024

Hi,
Iroko bench is already implemented

See: the tasks.py
Currently only the swahili subset is available, because we don't have correct translations for anchor words, that are required during the evaluation (e.g. answer/question).

If you are a native speaker in any of the languages we would love your help in improving that!
See the guide how to add a translations:

@dadelani
Copy link
Author

Great, thank you @clefourrier and @hynky1999 .

For the anchor words, maybe you can use google translate of the keywords. I can provide for Yoruba below:

Question: Ìbéèrè
Answer: Ìdáhùn

@NathanHB
Copy link
Member

NathanHB commented Oct 24, 2024

Great ! do you think you can open a PR for the evals and the keywords ? We will not use google translate as it can be unreliable.

As Clementine said, the documentation to add a translation is here:
https://github.com/huggingface/lighteval/wiki/Contributing-to-multilingual-evaluations

@clefourrier clefourrier added good first issue Good for newcomers help wanted Extra attention is needed labels Nov 8, 2024
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Labels
good first issue Good for newcomers help wanted Extra attention is needed new task
Projects
None yet
Development

No branches or pull requests

4 participants