Attention-based neural networks have become pervasive in many AI tasks. Despite their excellent algorithmic performance, the use of the attention mechanism and feed-forward network (FFN) demands excessive computational and memory resources, which often compromises their hardware performance. Although various sparse variants have been introduced, most approaches only focus on mitigating the quadratic scaling of attention on the algorithm level, without explicitly considering the efficiency of mapping their methods on real hardware designs. Furthermore, most efforts only focus on either the attention mechanism or the FFNs but without jointly optimizing both parts, causing most of the current designs to lack scalability when dealing with diffe...
Doctor of PhilosophyDepartment of Computer ScienceArslan MunirDeep neural networks (DNNs) have gaine...
The attention mechanism is the key to many state-of-the-art transformer-based models in Natural Lang...
The optical neural network (ONN) is a promising hardware platform for next-generation neurocomputing...
This repo contains the artifacts for our MICRO'22 paper titled "Adaptable Butterfly Accelerator for ...
The study of specialized accelerators tailored for neural networks is becoming a promising topic in ...
Spiking Neural Networks (SNNs) are bio-plausible models that hold great potential for realizing ener...
Overparameterized neural networks generalize well but are expensive to train. Ideally, one would lik...
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy cons...
© 2021 IEEE.The self-attention mechanism is rapidly emerging as one of the most important key primit...
Spiking Neural Networks (SNN) are an emerging type of biologically plausible and efficient Artificia...
Sparsity – the presence of many zero values – is a pervasive property of modern deep neural networks...
Deep neural networks virtually dominate the domain of most modern vision systems, providing high per...
Implementing embedded neural network processing at the edge requires efficient hardware acceleration...
Compiler frameworks are crucial for the widespread use of FPGA-based deep learning accelerators. The...
Long Short-Term Memory (LSTM) recurrent networks are frequently used for tasks involving time-sequen...
Doctor of PhilosophyDepartment of Computer ScienceArslan MunirDeep neural networks (DNNs) have gaine...
The attention mechanism is the key to many state-of-the-art transformer-based models in Natural Lang...
The optical neural network (ONN) is a promising hardware platform for next-generation neurocomputing...
This repo contains the artifacts for our MICRO'22 paper titled "Adaptable Butterfly Accelerator for ...
The study of specialized accelerators tailored for neural networks is becoming a promising topic in ...
Spiking Neural Networks (SNNs) are bio-plausible models that hold great potential for realizing ener...
Overparameterized neural networks generalize well but are expensive to train. Ideally, one would lik...
Spiking neural networks (SNNs) have achieved orders of magnitude improvement in terms of energy cons...
© 2021 IEEE.The self-attention mechanism is rapidly emerging as one of the most important key primit...
Spiking Neural Networks (SNN) are an emerging type of biologically plausible and efficient Artificia...
Sparsity – the presence of many zero values – is a pervasive property of modern deep neural networks...
Deep neural networks virtually dominate the domain of most modern vision systems, providing high per...
Implementing embedded neural network processing at the edge requires efficient hardware acceleration...
Compiler frameworks are crucial for the widespread use of FPGA-based deep learning accelerators. The...
Long Short-Term Memory (LSTM) recurrent networks are frequently used for tasks involving time-sequen...
Doctor of PhilosophyDepartment of Computer ScienceArslan MunirDeep neural networks (DNNs) have gaine...
The attention mechanism is the key to many state-of-the-art transformer-based models in Natural Lang...
The optical neural network (ONN) is a promising hardware platform for next-generation neurocomputing...