Call for Papers in the 1st Workshop on Scaling Up Multilingual Evaluation (SUMEval)
Massively Multilingual Language Models (MMLMs) are trained on around 100 languages of the world, however, most existing multilingual NLP benchmarks provide evaluation data in only a handful of these languages. The languages present in evaluation benchmarks are usually high-resource and largely belong to the Indo-European language family. This makes current multilingual evaluation unreliable and does not provide a full picture of the performance of MMLMs across the linguistic landscape. Although efforts are being made to create benchmarks that cover a larger variety of tasks, languages, and language families, it is unlikely that we will be able to build benchmarks covering all languages and tasks. Due to this, there is recent interest in alternate strategies for evaluating MMLMs, including performance prediction and Machine Translation of test data. We believe that this is an important yet relatively unexplored area of research that has the potential to make language technologies accessible to all.
*** Important Dates ***Workshop Paper Submission Deadline: August 25, 2022
Notification of Acceptance: September 25, 2022
Camera-ready papers due: October 10, 2022
Workshop Date: November 23, 2022
All deadlines are 11:59 pm UTC -12h ("anywhere on Earth").
*** Papers ***We invite submissions on topics that include, but are not limited to, the
following:
* Studies on scaling up multilingual evaluation
* Human evaluation of multilingual models
* Automated evaluation metrics for multilingual evaluation
* Studies on fairness and other aspects of evaluation
* Data sets, benchmarks or libraries for evaluating multi-lingual models
* Probing and analysis of multilingual models
For questions and comments regarding the workshop please contact the organizers at
sum...@microsoft.comPlease also follow the SUMEVal twitter account (
@sum_eval) for regular updates!
Looking forward to your submissions,
Best,
On behalf of the SUMEval 2022 Organizers,