In this paper we examine several combinations of classical N-gram language models with more advanced and well known techniques based on word similarity such as cache models and Latent Semantic Analysis. We compare the efficiency of these combined models to a model that combines N-grams with the recently proposed, state-of-the-art neural network-based continuous skip-gram. We discuss the strengths and weaknesses of each of these models, based on their predictive power of the Dutch language and find that a linear interpolation of a 3-gram, a cache model and a continuous skip-gram is capable of reducing perplexity by up to 18.63%, compared to a 3-gram baseline. This is three times the reduction achieved with a 5-gram.In addition, we investigat...
This study investigates the learning mechanisms underlying the acquisition of a dialect as a second ...
This thesis intends to compare the influence of lexical frequency with the influence of a constraint...
Languages are not uniform. Speakers of different language varieties use certain words differently - ...
© 2014 Pelemans et al.. In this paper we examine several combinations of classical N-gram language m...
Pelemans J., De Laet B., Van hamme H., Wambacq P., ''The effect of word similarity on N-gram languag...
Verwimp L., Pelemans J., Van hamme H., Wambacq P., ''Extending n-gram language models based on equiv...
Since the advent of deep learning, automatic speech recognition (ASR), like many other fields, has a...
© 2015 Lyan Verwimp, Joris Pelemans, Hugo Van hamme, Patrick Wambacq. The subject of this paper is t...
This article builds on computational tools to investigate the syntactic relationship between the hig...
International audienceThis paper describes an extension of the n-gram language model: the similar n-...
In recent years neural language models (LMs) have set state-of-the-art performance for several bench...
Conventional n-gram language models are well-established as powerful yet simple mechanisms for chara...
Abstract. This paper investigates a variety of statistical cache-based language models built upon th...
In this paper we investigate the differences in risk-averse behavior in translated versus non-transl...
Bilingual speakers store cognates from related languages close together in their mental lexicon. In ...
This study investigates the learning mechanisms underlying the acquisition of a dialect as a second ...
This thesis intends to compare the influence of lexical frequency with the influence of a constraint...
Languages are not uniform. Speakers of different language varieties use certain words differently - ...
© 2014 Pelemans et al.. In this paper we examine several combinations of classical N-gram language m...
Pelemans J., De Laet B., Van hamme H., Wambacq P., ''The effect of word similarity on N-gram languag...
Verwimp L., Pelemans J., Van hamme H., Wambacq P., ''Extending n-gram language models based on equiv...
Since the advent of deep learning, automatic speech recognition (ASR), like many other fields, has a...
© 2015 Lyan Verwimp, Joris Pelemans, Hugo Van hamme, Patrick Wambacq. The subject of this paper is t...
This article builds on computational tools to investigate the syntactic relationship between the hig...
International audienceThis paper describes an extension of the n-gram language model: the similar n-...
In recent years neural language models (LMs) have set state-of-the-art performance for several bench...
Conventional n-gram language models are well-established as powerful yet simple mechanisms for chara...
Abstract. This paper investigates a variety of statistical cache-based language models built upon th...
In this paper we investigate the differences in risk-averse behavior in translated versus non-transl...
Bilingual speakers store cognates from related languages close together in their mental lexicon. In ...
This study investigates the learning mechanisms underlying the acquisition of a dialect as a second ...
This thesis intends to compare the influence of lexical frequency with the influence of a constraint...
Languages are not uniform. Speakers of different language varieties use certain words differently - ...