In this thesis we introduce conditional neural language models based on log-bilinear and recurrent neural networks with applications to multimodal learning and natural language understanding. We first introduce a LSTM encoder for learning visual-semantic embeddings for ranking the relevance of text to images in a joint embedding space. Next we introduce three log-bilinear models for generating image descriptions that integrate both additive and multiplicative interactions. Beyond image conditioning, we describe a multiplicative conditional neural language model for learning distributed representations of attributes and meta data. Our model allows for contextual word relatedness comparisons through decompositions of a word embedding tensor. ...
This open access book provides an overview of the recent advances in representation learning theory,...
In the field of natural language processing (NLP), recent research has shown that deep neural networ...
The current generation of neural network-based natural language processing models excels at learning...
We introduce two multimodal neural lan-guage models: models of natural language that can be conditio...
Recurrent neural networks (RNN) have gained a reputation for producing state-of-the-art results on m...
Inspired by recent advances in multimodal learning and machine translation, we introduce an encoder-...
Convolutional Neural Networks (CNNs) have shown to yield very strong results in several Computer Vis...
We presented a learning model that generated natural language description of images. The model utili...
Recurrent neural networks (RNNs) are exceptionally good models of distributions over natural languag...
Recently a variety of LSTM-based conditional language models (LM) have been applied across a range o...
This thesis focuses on proposing and addressing various tasks in the field of vision and language, a...
We present novel methods for analyzing the activation patterns of recurrent neural networks from a l...
This thesis introduces the concept of an encoder-decoder neural network and develops architectures f...
We introduce (1) a novel neural network structure for bilingual modeling of sentence pairs that allo...
Language modeling is a crucial component in a wide range of applications including speech recognitio...
This open access book provides an overview of the recent advances in representation learning theory,...
In the field of natural language processing (NLP), recent research has shown that deep neural networ...
The current generation of neural network-based natural language processing models excels at learning...
We introduce two multimodal neural lan-guage models: models of natural language that can be conditio...
Recurrent neural networks (RNN) have gained a reputation for producing state-of-the-art results on m...
Inspired by recent advances in multimodal learning and machine translation, we introduce an encoder-...
Convolutional Neural Networks (CNNs) have shown to yield very strong results in several Computer Vis...
We presented a learning model that generated natural language description of images. The model utili...
Recurrent neural networks (RNNs) are exceptionally good models of distributions over natural languag...
Recently a variety of LSTM-based conditional language models (LM) have been applied across a range o...
This thesis focuses on proposing and addressing various tasks in the field of vision and language, a...
We present novel methods for analyzing the activation patterns of recurrent neural networks from a l...
This thesis introduces the concept of an encoder-decoder neural network and develops architectures f...
We introduce (1) a novel neural network structure for bilingual modeling of sentence pairs that allo...
Language modeling is a crucial component in a wide range of applications including speech recognitio...
This open access book provides an overview of the recent advances in representation learning theory,...
In the field of natural language processing (NLP), recent research has shown that deep neural networ...
The current generation of neural network-based natural language processing models excels at learning...