In the case of unwritten languages, acoustic models cannot be trained in the standard way, i.e., using speech and textual transcriptions. Recently, several methods have been proposed to learn speech representations using images, i.e., using visual grounding. Existing studies have focused on scene images. Here, we investigate whether fine-grained semantic information, reflecting the relationship between attributes and objects, can be learned from spoken language. To this end, a Fine-grained Semantic Embedding Network (FSEN) for learning semantic representations of spoken language grounded by fine-grained images is proposed. For training, we propose an efficient objective function, which includes a matching constraint, an adversarial objectiv...
In this paper, we explore neural network models that learn to associate segments of spoken audio cap...
An estimated half of the world’s languages do not have a written form, making it impossible for thes...
Abstract In this paper, we explore neural network models that learn to associate segments of spoken...
Visually grounded speech representation learning has shown to be useful in the field of speech repre...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Comp...
Text-based technologies, such as text translation from one language to another, and image captioning...
We present a visually grounded model of speech perception which projects spoken utterances and image...
This paper explores the possibility to learn a semantically-relevant lexicon from images and speech ...
<p>This paper explores the possibility to learn a semantically-relevant lexicon from images and spee...
Humans learn language by interaction with their environment and listening to other humans. It should...
A widespread approach to processing spoken language is to first automatically transcribe it into tex...
We investigated word recognition in a Visually Grounded Speech model. The model has been trained on ...
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Com...
Visual Grounding (VG) is a task of locating a specific object in an image semantically matching a gi...
International audienceThe language acquisition literature shows that children do not build their lex...
In this paper, we explore neural network models that learn to associate segments of spoken audio cap...
An estimated half of the world’s languages do not have a written form, making it impossible for thes...
Abstract In this paper, we explore neural network models that learn to associate segments of spoken...
Visually grounded speech representation learning has shown to be useful in the field of speech repre...
Thesis: Ph. D., Massachusetts Institute of Technology, Department of Electrical Engineering and Comp...
Text-based technologies, such as text translation from one language to another, and image captioning...
We present a visually grounded model of speech perception which projects spoken utterances and image...
This paper explores the possibility to learn a semantically-relevant lexicon from images and speech ...
<p>This paper explores the possibility to learn a semantically-relevant lexicon from images and spee...
Humans learn language by interaction with their environment and listening to other humans. It should...
A widespread approach to processing spoken language is to first automatically transcribe it into tex...
We investigated word recognition in a Visually Grounded Speech model. The model has been trained on ...
Thesis: M. Eng., Massachusetts Institute of Technology, Department of Electrical Engineering and Com...
Visual Grounding (VG) is a task of locating a specific object in an image semantically matching a gi...
International audienceThe language acquisition literature shows that children do not build their lex...
In this paper, we explore neural network models that learn to associate segments of spoken audio cap...
An estimated half of the world’s languages do not have a written form, making it impossible for thes...
Abstract In this paper, we explore neural network models that learn to associate segments of spoken...