The Transformer architecture is widely used for machine translation tasks. However, its resource-intensive nature makes it challenging to implement on constrained embedded devices, particularly where available hardware resources can vary at run-time. We propose a dynamic machine translation model that scales the Transformer architecture based on the available resources at any particular time. The proposed approach, 'Dynamic-HAT', uses a HAT SuperTransformer as the backbone to search for SubTransformers with different accuracy-latency trade-offs at design time. The optimal SubTransformers are sampled from the SuperTransformer at run-time, depending on latency constraints. The Dynamic-HAT is tested on the Jetson Nano and the approach uses inh...
Machine translation has received significant attention in the field of natural language processing n...
Retrosynthesis is the task of building a molecule from smaller precursor molecules. As shown in prev...
Dynamic binary translation is the process of translating and optimizing executable code for one mach...
This dataset supports the publication: 'Dynamic Transformer for Efficient Machine Translation on...
With the recent developments in the field of Natural Language Processing, there has been a rise in t...
With the recent developments in the field of Natural Language Processing, there has been a rise in t...
This article describes our experiments in neural machine translation using the recent Tensor2Tensor ...
Pre-trained language models received extensive attention in recent years. However, it is still chall...
Given a large Transformer model, how can we obtain a small and computationally efficient model which...
The Transformer architecture is ubiquitously used as the building block of large-scale autoregressiv...
Transformer networks have emerged as the state-of-the-art approach for natural language processing t...
Machine translation is the discipline concerned with developing automated tools for translating fro...
Given a large Transformer model, how can we obtain a small and computationally efficient model which...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
The topic of transformers is rapidly emerging as one of the most important key primitives in neural ...
Machine translation has received significant attention in the field of natural language processing n...
Retrosynthesis is the task of building a molecule from smaller precursor molecules. As shown in prev...
Dynamic binary translation is the process of translating and optimizing executable code for one mach...
This dataset supports the publication: 'Dynamic Transformer for Efficient Machine Translation on...
With the recent developments in the field of Natural Language Processing, there has been a rise in t...
With the recent developments in the field of Natural Language Processing, there has been a rise in t...
This article describes our experiments in neural machine translation using the recent Tensor2Tensor ...
Pre-trained language models received extensive attention in recent years. However, it is still chall...
Given a large Transformer model, how can we obtain a small and computationally efficient model which...
The Transformer architecture is ubiquitously used as the building block of large-scale autoregressiv...
Transformer networks have emerged as the state-of-the-art approach for natural language processing t...
Machine translation is the discipline concerned with developing automated tools for translating fro...
Given a large Transformer model, how can we obtain a small and computationally efficient model which...
Humans benefit from communication but suffer from language barriers. Machine translation (MT) aims t...
The topic of transformers is rapidly emerging as one of the most important key primitives in neural ...
Machine translation has received significant attention in the field of natural language processing n...
Retrosynthesis is the task of building a molecule from smaller precursor molecules. As shown in prev...
Dynamic binary translation is the process of translating and optimizing executable code for one mach...