Slot and Intent Detection Resources for Bavarian and Lithuanian: Assessing Translations vs Natural Queries to Digital Assistants
-
We present NaBaLiSID (Natural Lithuanian and Bavarian SID), which contains new slot and intent detection evaluation datasets for Bavarian and Lithuanian, created by manual translation and adopting the translation and annotation schemes used by van der Goot et al. (2021) for xSID.
-
We collect natural datasets of utterances from native speakers to be able to evaluate on more realistic data.
-
For Bavarian, we further present translations of a part of the large MASSIVE (FitzGerald et al., 2023) dataset to Bavarian to evaluate the effect of transferring to a low-resource language without orthography in a cross-datasets setting.
-
We evaluate the performance of cross-lingual language models on our translated and native data with the MaChAmp toolkit (van der Goot et al., 2021), to gauge the effect of having natural utterances versus translations for SID.
-
data
: annotated NaLiBaSID datasetsaccess the data with the password
MaiNLP
- de-ba.test & .valid: Bavarian xSID translations de-ba
- lt.test & .valid: Lithuanian xSID translations lt
- de-ba.MAS.test + .valid: Bavarian MASSIVE translation MAS:de-ba
- de-ba.xMAS.test + .valid: Bavarian xSID + MASSIVE translations xMAS:de-ba
- de-ba.nat: Natural Bavarian data nat:de-ba
- lt.nat: Natural Lithuanian data nat:lt
-
configs
:- MaChAmp dataset and parameter configs used for producing the results presented in the paper
-
scripts
:- Run setup and training of all models with
scripts/00_runAll.sh
(training may lead to a long duration) - Alternatively run each step individually with the provided scripts
- Let the models predict on the test data with MaChAmp's
predict.py
- Script for metric calculations (
nluEval.py
) is provided by van der Goot et al. (2021).
- Run setup and training of all models with
-
predictions
:- Output predictions and model scores. Produced with MaChAmp's
predict.py
- access the data with the password
MaiNLP
whole_data
contains the natural datasets that are not split into test and development setstest_data
andvalid_data
also include the predictions on the English and German xSID datasets
- Output predictions and model scores. Produced with MaChAmp's
If you use the data and/or code in this repository, please cite the following:
@inproceedings{Winkler2024,
title = "Slot and Intent Detection Resources for {B}avarian and {L}ithuanian: Assessing Translations vs Natural Queries to Digital Assistants",
author = "Winkler, Miriam and Juozapaityte, Virginija and van der Goot, Rob and Plank, Barbara",
booktitle = "Proceedings of The 2024 Joint International Conference on Computational Linguistics, Language Resources and Evaluation",
year = "2024",
publisher = "Association for Computational Linguistics",
}
@inproceedings{van-der-goot-etal-2020-cross,
title={From Masked-Language Modeling to Translation: Non-{E}nglish Auxiliary Tasks Improve Zero-shot Spoken Language Understanding},
author={van der Goot, Rob and Sharaf, Ibrahim and Imankulova, Aizhan and {\"U}st{\"u}n, Ahmet and Stepanovic, Marija and Ramponi, Alan and Khairunnisa, Siti Oryza and Komachi, Mamoru and Plank, Barbara},
booktitle = "Proceedings of the 2021 Conference of the North {A}merican Chapter of the Association for Computational Linguistics: Human Language Technologies, Volume 1 (Long and Short Papers)",
year = "2021",
address = "Mexico City, Mexico",
publisher = "Association for Computational Linguistics"
}
For more languages, please refer to the xSID GitHub repository.
This work is supported by ERC Consolidator Grant DIALECT no. 101043235.