resource languages have significantly lower attack success rate, which suggests that the cross-lingual vulnerability mainly applies to low-resource languages. Previously, limited training on low-resource languages primarily affects speakers of those languages, causing technological disparities. However, our ...
内容提示: Meeting the Needs of Low-Resource Languages:The Value of Automatic Alignments via Pretrained ModelsAbteen Ebrahimi ♦ Arya D. McCarthy ∇ Arturo Oncevay ♥Luis Chiruzzo 4 John E. Ortega Ω Gustavo A. Giménez-Lugo ♣Rolando Coto-Solano φ Katharina Kann ♦♦ University ...
Low Resource Languages Resources for conservation, development, and documentation of low resource (human) languages. According to some estimates, half of the 7,000~ currently spoken languages are expected to become extinct this century. However, there is a lot of work by academics, independent scho...
Intuitively, in the machine translation paradigm, especially in low-resource languages scenario, many DA methods have appeared. The most commonly used methods are building pseudocorpus by randomly sampling, omitting, or replacing some words in the text. However, previous approaches hardly guarantee the...
,P Mitra - ACM Transactions on Asian and Low-Resource Language Information Processing 被引量: 0发表: 2023年 Query-by-Example Spoken Term Detection for Low-resource Languages In this thesis, we consider the problem of query-by-example (QbyE) spoken term detection (STD) for low-resource ...
“In general, multilingual language models struggle with languages written in non-Latin scripts, language isolates, and families of languages less connected to those of high resource languages,” they wrote. “This threatens to create a poor-get-poorer dynamic for languages that are only similar to...
1. Building language models for low-resource languages. 2. Adapting/extending existing language models/large language models for low-resource languages. 3. Corpora creation and curation technologies for training language models/large language models for low-resource languages. ...
Language models struggle in generating correct code for low resource programming languages, since these are underrepresented in training data. Popular approaches use either examples or documentation to improve the performance of these models. Instead of considering the indep...
Recently, very large language models (LLMs) have shown exceptional performance on several English NLP tasks with just in-context learning (ICL), but their utility in other languages is still underexplored. We investigate their effectiveness for NLP tasks in low-resource languages (LRLs), especially...
Multi-lingual language models (LM), such as mBERT, XLM-R, mT5, mBART, have been remarkably successful in enabling natural language tasks in low-resource languages through cross-lingual transfer from high-resource ones. In this work, we try to better understand how such models, specifically mT5...