Recent work demonstrates the potential of multilingual pretraining of creating one model that can be used for various tasks in different languages. Previous work in multilingual pretraining has demonstrated that
Zero Shotmeans that the Multilingual BERT system was fine-tuned on English MultiNLI, and then evaluated on the foreign language XNLI test. In this case, machine translation was not involved at all in either the pre-training or fine-tuning. ...
Urdu Vietnamese As per recommendations in xlm-roberta-large-xnli, for English-only classification, you might want to check out: bart-large-mnli a distilled bart MNLI model. Zero-shot example: The model retains its text-to-text characteristic after fine-tuning. This means that our expected out...
Urdu urd 1 Uzbek uzb 1 Venda ven 1 Vietnamese vie 1 Wollof wol 1 Yambassa mis 1 Yiddish ydd 1 Zambia bem 1 Unknown Unknown 10 Table 10. Comparison of Eru Dishes in WWD and WC. The table compares the representation of the traditional dish Eru in the WWD and WC datasets. Note how...