The thesis deals with language modelling for German. The main concerns are the specifics of German language that are troublesome for standard n-gram models. First the statistical methods of language modelling are described and language phenomena of German are explained. Following that suggests own variants of n-gram language models with an aim to improve these problems. The models themselves are trained using the standard n-gram methods as well as using the method of maximum entropy with n-gram features. Both possibilities are compared using corelation metrics of hand-evaluated fluency of sentences and automatic evaluation - the perplexity. Also, the computation requirements are compared. Next, the thesis presents a set of own features that...
Abstract: In this paper we present a synthesis of the theoretical fundamentals and some practical as...
The paper presents a large-scale computational subcategorisation lexicon for several thousand German...
Language Modeling is one of the most important subfields of modern Natural Language Processing (NLP)....
The thesis deals with language modelling for German. The main concerns are the specifics of German l...
In this work we address the challenge of augmenting n-gram language models according to prior lingui...
Statistical n-gram language modeling is used in many domains like speech recognition, language ident...
© 2015 Lyan Verwimp, Joris Pelemans, Hugo Van hamme, Patrick Wambacq. The subject of this paper is t...
This PhD thesis studies the overall effect of statistical language modeling on perplexity and word e...
Most Natural Language Processing (NLP) applications focus on standardized, written language varietie...
Cross-language comparison can show where German structures diverge from and converge with those of o...
It seems obvious that a successful model of natural language would incorporate a great deal of both ...
This work deals with the application that uses the machine-learning methods for the automatic langua...
This report documents the program and the outcomes of Dagstuhl Seminar 14061 "Statistical Techniques...
We present a tutorial introduction to n-gram models for language modeling and survey the most widely...
A new language model is presented which incorporates local N-gram dependencies with two important so...
Abstract: In this paper we present a synthesis of the theoretical fundamentals and some practical as...
The paper presents a large-scale computational subcategorisation lexicon for several thousand German...
Language Modeling is one of the most important subfields of modern Natural Language Processing (NLP)....
The thesis deals with language modelling for German. The main concerns are the specifics of German l...
In this work we address the challenge of augmenting n-gram language models according to prior lingui...
Statistical n-gram language modeling is used in many domains like speech recognition, language ident...
© 2015 Lyan Verwimp, Joris Pelemans, Hugo Van hamme, Patrick Wambacq. The subject of this paper is t...
This PhD thesis studies the overall effect of statistical language modeling on perplexity and word e...
Most Natural Language Processing (NLP) applications focus on standardized, written language varietie...
Cross-language comparison can show where German structures diverge from and converge with those of o...
It seems obvious that a successful model of natural language would incorporate a great deal of both ...
This work deals with the application that uses the machine-learning methods for the automatic langua...
This report documents the program and the outcomes of Dagstuhl Seminar 14061 "Statistical Techniques...
We present a tutorial introduction to n-gram models for language modeling and survey the most widely...
A new language model is presented which incorporates local N-gram dependencies with two important so...
Abstract: In this paper we present a synthesis of the theoretical fundamentals and some practical as...
The paper presents a large-scale computational subcategorisation lexicon for several thousand German...
Language Modeling is one of the most important subfields of modern Natural Language Processing (NLP)....