Models based around the transformer architecture have become one of the most prominent for solving a multitude of natural language processing (NLP)tasks since its introduction in 2017. However, much research related to the transformer model has focused primarily on achieving high performance and many problems remain unsolved. Two of the most prominent currently are the lack of high performing non-English pre-trained models, and the limited number of words most trained models can incorporate for their context. Solving these problems would make NLP models more suitable for real-world applications, improving information retrieval, reading comprehension, and more. All previous research has focused on incorporating long-context for English langu...
The complexity for being at the forefront regarding information retrieval systems are constantly inc...
Short-text matching is a fundamental task in many important NLP applications such as question answer...
Thesis (Ph.D.)--University of Washington, 2022Modern NLP systems have been highly successful at a wi...
Models based around the transformer architecture have become one of the most prominent for solving a...
The main goal behind state-of-the-art pretrained multilingual models such as multilingual BERT and X...
Prior work on multilingual question answering has mostly focused on using large multilingual pre-tra...
While several benefits were realized for multilingual vision-language pretrained models, recent benc...
The research of open-domain, knowledge-grounded dialogue systems has been advancing rapidly due to t...
Cross-lingual Machine Reading Comprehension (xMRC) is a challenging task due to the lack of training...
While recent language models have the ability to take long contexts as input, relatively little is k...
In the modern era of deep learning, developing natural language processing (NLP) systems require lar...
Question Answering systems are greatly sought after in many areas of industry. Unfortunately, as mos...
Large pre-trained multilingual models such as mBERT and XLM-R enabled effective cross-lingual zero-s...
International audienceSupervised deep learning-based approaches have been applied to task-oriented d...
Large pre-trained masked language models have become state-of-the-art solutions for many NLP problem...
The complexity for being at the forefront regarding information retrieval systems are constantly inc...
Short-text matching is a fundamental task in many important NLP applications such as question answer...
Thesis (Ph.D.)--University of Washington, 2022Modern NLP systems have been highly successful at a wi...
Models based around the transformer architecture have become one of the most prominent for solving a...
The main goal behind state-of-the-art pretrained multilingual models such as multilingual BERT and X...
Prior work on multilingual question answering has mostly focused on using large multilingual pre-tra...
While several benefits were realized for multilingual vision-language pretrained models, recent benc...
The research of open-domain, knowledge-grounded dialogue systems has been advancing rapidly due to t...
Cross-lingual Machine Reading Comprehension (xMRC) is a challenging task due to the lack of training...
While recent language models have the ability to take long contexts as input, relatively little is k...
In the modern era of deep learning, developing natural language processing (NLP) systems require lar...
Question Answering systems are greatly sought after in many areas of industry. Unfortunately, as mos...
Large pre-trained multilingual models such as mBERT and XLM-R enabled effective cross-lingual zero-s...
International audienceSupervised deep learning-based approaches have been applied to task-oriented d...
Large pre-trained masked language models have become state-of-the-art solutions for many NLP problem...
The complexity for being at the forefront regarding information retrieval systems are constantly inc...
Short-text matching is a fundamental task in many important NLP applications such as question answer...
Thesis (Ph.D.)--University of Washington, 2022Modern NLP systems have been highly successful at a wi...