With the increasing availability of large datasets machine learning techniques are be-coming an increasingly attractive alternative to expert-designed approaches to solving complex problems in domains where data is abundant. In this thesis we introduce several models for large sparse discrete datasets. Our approach, which is based on probabilistic models that use distributed representations to alleviate the effects of data sparsity, is applied to statistical language modelling and collaborative filtering. We introduce three probabilistic language models that represent words using learned real-valued vectors. Two of the models are based on the Restricted Boltzmann Machine (RBM) architecture while the third one is a simple deterministic model...
Grammar-based natural language processing has reached a level where it can `understand' language to ...
Multi-task learning seeks to improve the generalization performance by sharing common information am...
This paper reports on the benefits of largescale statistical language modeling in machine translatio...
With the increasing availability of large datasets machine learning techniques are becoming an incr...
Statistical language models estimate the probability of a word occurring in a given context. The mos...
The restricted Boltzmann machine (RBM) is a flexible tool for modeling complex data, however there h...
Finding the right representations for words is critical for building accurate NLP systems when domai...
Language models are probability distributions over a set of unilingual natural language text used in...
Combined with neural language models, distributed word representations achieve significant advantage...
In the field of Natural Language Processing, supervised machine learning is commonly used to solve c...
Finding the right representations for words is critical for building accurate NLP systems when domai...
Vector-space distributed representations of words can capture syntactic and semantic regularities in...
In this paper, we consider learning dictionary models over a network of agents, where each agent is ...
This thesis, which is organized in two independent parts, presents work on distributional semantics ...
In spite of their superior performance, neural probabilistic language models (NPLMs) remain far less...
Grammar-based natural language processing has reached a level where it can `understand' language to ...
Multi-task learning seeks to improve the generalization performance by sharing common information am...
This paper reports on the benefits of largescale statistical language modeling in machine translatio...
With the increasing availability of large datasets machine learning techniques are becoming an incr...
Statistical language models estimate the probability of a word occurring in a given context. The mos...
The restricted Boltzmann machine (RBM) is a flexible tool for modeling complex data, however there h...
Finding the right representations for words is critical for building accurate NLP systems when domai...
Language models are probability distributions over a set of unilingual natural language text used in...
Combined with neural language models, distributed word representations achieve significant advantage...
In the field of Natural Language Processing, supervised machine learning is commonly used to solve c...
Finding the right representations for words is critical for building accurate NLP systems when domai...
Vector-space distributed representations of words can capture syntactic and semantic regularities in...
In this paper, we consider learning dictionary models over a network of agents, where each agent is ...
This thesis, which is organized in two independent parts, presents work on distributional semantics ...
In spite of their superior performance, neural probabilistic language models (NPLMs) remain far less...
Grammar-based natural language processing has reached a level where it can `understand' language to ...
Multi-task learning seeks to improve the generalization performance by sharing common information am...
This paper reports on the benefits of largescale statistical language modeling in machine translatio...