Semantic representation learning for sentences is an important and well-studied problem in NLP. The current trend for this task involves training a Transformer-based sentence encoder through a contrastive objective with text, i.e., clustering sentences with semantically similar meanings and scattering others. In this work, we find the performance of Transformer models as sentence encoders can be improved by training with multi-modal multi-task losses, using unpaired examples from another modality (e.g., sentences and unrelated image/audio data). In particular, besides learning by the contrastive loss on text, our model clusters examples from a non-linguistic domain (e.g., visual/audio) with a similar contrastive loss at the same time. The r...
Contrastive learning has become a new paradigm for unsupervised sentence embeddings. Previous studie...
We propose PromptBERT, a novel contrastive learning method for learning better sentence representati...
This study addresses the question whether visually grounded speech recognition (VGS) models learn to...
The question of what kinds of linguistic information are encoded in different layers of Transformer-...
Several prior studies have suggested that word frequency biases can cause the Bert model to learn in...
Contrastive learning has been demonstrated to be effective in enhancing pre-trained language models ...
Though offering amazing contextualized token-level representations, current pre-trained language mod...
Unsupervised sentence representation learning is a fundamental problem in natural language processin...
Pretrained Masked Language Models (MLMs) have revolutionised NLP in recent years. However, previous ...
Current approaches to learning semantic representations of sentences often use prior word-level know...
Many modern NLP systems rely on word embeddings, previously trained in an unsupervised manner on lar...
The field of Natural Language Processing (NLP) has progressed rapidly in recent years due to the evo...
Natural language processing (NLP) is one of the most important technologies of the information age. ...
Contains fulltext : 235108.pdf (Publisher’s version ) (Open Access)Interspeech 202...
Contrastive learning-based methods, such as unsup-SimCSE, have achieved state-of-the-art (SOTA) perf...
Contrastive learning has become a new paradigm for unsupervised sentence embeddings. Previous studie...
We propose PromptBERT, a novel contrastive learning method for learning better sentence representati...
This study addresses the question whether visually grounded speech recognition (VGS) models learn to...
The question of what kinds of linguistic information are encoded in different layers of Transformer-...
Several prior studies have suggested that word frequency biases can cause the Bert model to learn in...
Contrastive learning has been demonstrated to be effective in enhancing pre-trained language models ...
Though offering amazing contextualized token-level representations, current pre-trained language mod...
Unsupervised sentence representation learning is a fundamental problem in natural language processin...
Pretrained Masked Language Models (MLMs) have revolutionised NLP in recent years. However, previous ...
Current approaches to learning semantic representations of sentences often use prior word-level know...
Many modern NLP systems rely on word embeddings, previously trained in an unsupervised manner on lar...
The field of Natural Language Processing (NLP) has progressed rapidly in recent years due to the evo...
Natural language processing (NLP) is one of the most important technologies of the information age. ...
Contains fulltext : 235108.pdf (Publisher’s version ) (Open Access)Interspeech 202...
Contrastive learning-based methods, such as unsup-SimCSE, have achieved state-of-the-art (SOTA) perf...
Contrastive learning has become a new paradigm for unsupervised sentence embeddings. Previous studie...
We propose PromptBERT, a novel contrastive learning method for learning better sentence representati...
This study addresses the question whether visually grounded speech recognition (VGS) models learn to...