Conformer has proven to be effective in many speech processing tasks. It combines the benefits of extracting local dependencies using convolutions and global dependencies using self-attention. Inspired by this, we propose a more flexible, interpretable and customizable encoder alternative, Branchformer, with parallel branches for modeling various ranged dependencies in end-to-end speech processing. In each encoder layer, one branch employs self-attention or its variant to capture long-range dependencies, while the other branch utilizes an MLP module with convolutional gating (cgMLP) to extract local relationships. We conduct experiments on several speech recognition and spoken language understanding benchmarks. Results show that our model o...
Currently, there are mainly three Transformer encoder based streaming End to End (E2E) Automatic Spe...
Self-attention mechanisms model long-range context by using pairwise attention between all input tok...
This paper presents an in-depth study on a Sequentially Sampled Chunk Conformer, SSC-Conformer, for ...
International audienceThe recently proposed Conformer architecture has shown state-of-the-art perfor...
This study addresses robust automatic speech recognition (ASR) by introducing a Conformer-based acou...
Convolutional neural networks (CNN) and Transformer have wildly succeeded in multimedia applications...
Owing to the loss of effective information and incomplete feature extraction caused by the convoluti...
Model fine-tuning and adaptation have become a common approach for model specialization for downstre...
In this paper, we present Multi-scale Feature Aggregation Conformer (MFA-Conformer), an easy-to-impl...
Optimization of modern ASR architectures is among the highest priority tasks since it saves many com...
Speech processing is highly incremental. It is widely accepted that human listeners continuously use...
Traditionally, research in automated speech recognition has focused on local-first encoding of audio...
While transformers and their variant conformers show promising performance in speech recognition, th...
Large-scale speech self-supervised learning (SSL) has emerged to the main field of speech processing...
Recent progress in NLP witnessed the development of large-scale pre-trained language models (GPT, BE...
Currently, there are mainly three Transformer encoder based streaming End to End (E2E) Automatic Spe...
Self-attention mechanisms model long-range context by using pairwise attention between all input tok...
This paper presents an in-depth study on a Sequentially Sampled Chunk Conformer, SSC-Conformer, for ...
International audienceThe recently proposed Conformer architecture has shown state-of-the-art perfor...
This study addresses robust automatic speech recognition (ASR) by introducing a Conformer-based acou...
Convolutional neural networks (CNN) and Transformer have wildly succeeded in multimedia applications...
Owing to the loss of effective information and incomplete feature extraction caused by the convoluti...
Model fine-tuning and adaptation have become a common approach for model specialization for downstre...
In this paper, we present Multi-scale Feature Aggregation Conformer (MFA-Conformer), an easy-to-impl...
Optimization of modern ASR architectures is among the highest priority tasks since it saves many com...
Speech processing is highly incremental. It is widely accepted that human listeners continuously use...
Traditionally, research in automated speech recognition has focused on local-first encoding of audio...
While transformers and their variant conformers show promising performance in speech recognition, th...
Large-scale speech self-supervised learning (SSL) has emerged to the main field of speech processing...
Recent progress in NLP witnessed the development of large-scale pre-trained language models (GPT, BE...
Currently, there are mainly three Transformer encoder based streaming End to End (E2E) Automatic Spe...
Self-attention mechanisms model long-range context by using pairwise attention between all input tok...
This paper presents an in-depth study on a Sequentially Sampled Chunk Conformer, SSC-Conformer, for ...