Hi Samuel,
Thank you for your question. We've decided on a format for the QA task; it will be JSONL files with each line looking something like this:
{"doc_id": 0, "question_id": "A1.1.H01", "pred": 1}
This is an example from the development set for Sorbian.
There should be one output file per test set input file. The required fields are 'doc_id' (int), 'question_id' (str) for Sorbian, and 'pred' (int). The Ukrainian QA doesn't have the 'question_id' field so this can be None. For Sorbian QA the 'question_id' is the uniquely identifying one even across different files. You may add other fields to your output, e.g., the question and possible answers for manual inspection, but we will evaluate based on the required fields.
For MT, we'll be coordinating with the General shared task organisers about the expected format.
We additionally plan to provide a helper script for converting the lm-evalharness output to the expected submission format.