Thesis (Master's)--University of Washington, 2020Understanding language depending on the context of its usage has always been one of thecore goals of natural language processing. Recently, contextual word representations created by language models like ELMo, BERT, ELECTRA, and RoBERTA have provided robust representations of natural language which serve as the language understanding component for a diverse range of downstream tasks like information retrieval, question answering, and information extraction. Curriculum learning is a method that employs a structured training regime instead of the traditional random sampling. Research areas like computer vision and machine translation have used curriculum learning methods in model training to im...
After becoming familiar with preparing text data in different formats and training different algorit...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
In this paper, we review recent progress in the field of machine learning and examine its implicatio...
We describe our team's contribution to the STRICT-SMALL track of the BabyLM Challenge. The challenge...
Transfer learning is to apply knowledge or patterns learned in a particular field or task to differe...
The nature and amount of information needed for learning a natural language, and the underlying mech...
In the last few decades, text mining has been used to extract knowledge from free texts. Applying ne...
Unsupervised learning text representations aims at converting natural languages into vector represen...
In the last few decades, text mining has been used to extract knowledge from free texts. Applying ne...
Language acquisition by children and machines is remarkable. Yet while children learn from hearing a...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
How to properly represent language is a crucial and fundamental problem in Natural Language Processi...
The current generation of neural network-based natural language processing models excels at learning...
These improvements open many possibilities in solving Natural Language Processing downstream tasks. ...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, F...
After becoming familiar with preparing text data in different formats and training different algorit...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
In this paper, we review recent progress in the field of machine learning and examine its implicatio...
We describe our team's contribution to the STRICT-SMALL track of the BabyLM Challenge. The challenge...
Transfer learning is to apply knowledge or patterns learned in a particular field or task to differe...
The nature and amount of information needed for learning a natural language, and the underlying mech...
In the last few decades, text mining has been used to extract knowledge from free texts. Applying ne...
Unsupervised learning text representations aims at converting natural languages into vector represen...
In the last few decades, text mining has been used to extract knowledge from free texts. Applying ne...
Language acquisition by children and machines is remarkable. Yet while children learn from hearing a...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
How to properly represent language is a crucial and fundamental problem in Natural Language Processi...
The current generation of neural network-based natural language processing models excels at learning...
These improvements open many possibilities in solving Natural Language Processing downstream tasks. ...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Brain and Cognitive Sciences, F...
After becoming familiar with preparing text data in different formats and training different algorit...
Currently, N-gram models are the most common and widely used models for statistical language modelin...
In this paper, we review recent progress in the field of machine learning and examine its implicatio...