Replies: 2 comments 11 replies
-
Great point! My view is that we should support them in the code & let anyone add them (with points). For the leaderboard later, I think we could have one "crosslingual" tab in addition to per-language tabs. Note that for STS we already have crosslingual datasets in the code & LB, but currently they are just in the |
Beta Was this translation helpful? Give feedback.
-
English->XX or XX->English is already a good base for cross-lingual tasks. STS supports this, but it would be nice to extend it to Retrieval, Reranking and maybe Summarization (not sure about existing datasets here) tasks. For XX -> YY (where XX != English), mlqa is also a good resource. @orionw should we just add new tasks inheriting from the current AbsTasks, or should we think about adding someting to point that it is a cross-lingual task (inheriting from |
Beta Was this translation helpful? Give feedback.
-
(started from a discussion in #347 on cross-lingual datasets with @KennethEnevoldsen and @izhx. Also tagging @Muennighoff)
Should we include cross-lingual datasets in MMTEB (such as xPCA, xQA, xOR-TyDiQA)?
Where I'm located (near Washington DC) people care a lot about cross-lingual but only for English->XX (where we have an English query and we're looking for documents in other languages). There are a decent amount of resources for these English-centric cross-lingual tasks, as listed above.
However, given that this benchmark is a worldwide effort, it might not make sense to focus on only En->XX or XX->En. On the other hand, given the number potential cross-lingual categories it would difficult to get a thorough benchmark for the cross-product of languages, as I doubt that there are existing resources for many of these pairs. However, we could include what we can find in a separate cross-lingual category.
What are people's thoughts on this? I might lean towards not supporting it in the initial version of MMTEB, but I don't have a strong preference.
Beta Was this translation helpful? Give feedback.
All reactions