We present a systematic study and comprehensive evaluation of large language models for automatic multilingual readability assessment. In particular, we construct ReadMe++, a multilingual multi-domain dataset with human annotations of 9757 sentences in Arabic, English, French, Hindi, and Russian collected from 112 different data sources. ReadMe++ offers more domain and language diversity than existing readability datasets, making it ideal for benchmarking multilingual and non-English language models (including mBERT, XLM-R, mT5, Llama-2, GPT-4, etc.) in the supervised, unsupervised, and few-shot prompting settings. Our experiments reveal that models fine-tuned on ReadMe++ outperform those trained on single-domain datasets, showcasing superi...
We release a new benchmark for Automated Readability Assessment (ARA) of texts in Spanish. We combin...
Although recent Massively Multilingual Language Models (MMLMs) like mBERT and XLMR support around 10...
Vikidia.org is a children's encyclopedia, with content targeting 8-13 year old children, in several ...
We present Belebele, a multiple-choice machine reading comprehension (MRC) dataset spanning 122 lang...
Recent benchmarks for Large Language Models (LLMs) have mostly focused on application-driven tasks s...
Readability refers to the ease with which a reader can understand a text. Automatic readability asse...
Current research on automatic readability assessment (ARA) has focused on improving the performance ...
Large Language Models (LLMs) have demonstrated impressive performance on Natural Language Processing...
In recent years, the main focus of research on automatic readability assessment (ARA) has shifted to...
The main limiting factor in the development of robust multilingual dialogue evaluation metrics is th...
Multilingual evaluation benchmarks usually contain limited high-resource languages and do not test m...
How cross-linguistically applicable are NLP models, specifically language models? A fair comparison ...
We introduce MADLAD-400, a manually audited, general domain 3T token monolingual dataset based on Co...
Automatic readability assessment is considered as a challenging task in NLP due to its high degree o...
ABSTRACT. Improved readability ratings for second-language readers could have a huge impact in areas...
We release a new benchmark for Automated Readability Assessment (ARA) of texts in Spanish. We combin...
Although recent Massively Multilingual Language Models (MMLMs) like mBERT and XLMR support around 10...
Vikidia.org is a children's encyclopedia, with content targeting 8-13 year old children, in several ...
We present Belebele, a multiple-choice machine reading comprehension (MRC) dataset spanning 122 lang...
Recent benchmarks for Large Language Models (LLMs) have mostly focused on application-driven tasks s...
Readability refers to the ease with which a reader can understand a text. Automatic readability asse...
Current research on automatic readability assessment (ARA) has focused on improving the performance ...
Large Language Models (LLMs) have demonstrated impressive performance on Natural Language Processing...
In recent years, the main focus of research on automatic readability assessment (ARA) has shifted to...
The main limiting factor in the development of robust multilingual dialogue evaluation metrics is th...
Multilingual evaluation benchmarks usually contain limited high-resource languages and do not test m...
How cross-linguistically applicable are NLP models, specifically language models? A fair comparison ...
We introduce MADLAD-400, a manually audited, general domain 3T token monolingual dataset based on Co...
Automatic readability assessment is considered as a challenging task in NLP due to its high degree o...
ABSTRACT. Improved readability ratings for second-language readers could have a huge impact in areas...
We release a new benchmark for Automated Readability Assessment (ARA) of texts in Spanish. We combin...
Although recent Massively Multilingual Language Models (MMLMs) like mBERT and XLMR support around 10...
Vikidia.org is a children's encyclopedia, with content targeting 8-13 year old children, in several ...