The field of Natural Language Processing (NLP) has been undergoing a revolution in recent years. Large-scale language models (LLMs), most notably a series of Generative Pre-trained Transformers (GPTs), exceeded all expectations in benchmark scenarios and real life applications such as text generation, translation, question-answering and summarization. The engine of the NLP revolution is the so-called attention mechanism, which now allows to process longer sentences without 'forgetting' important words. This mechanism is implemented in form of a series of matrix products and lends itself to intense parallelization. The pre-training of transformers requires great computational resources and is one example of the increasing AI workload of larg...
Neural machine translation (NMT) is an approach to machine translation (MT) that uses deep learning ...
This report documents the program and the outcomes of Dagstuhl Seminar 22232 "Efficient and Equitabl...
Transformer-based neural models are used in many AI applications. Training these models is expensive...
This report documents the program and the outcomes of Dagstuhl Seminar 22232 “Efficient and Equitabl...
Artificial neural networks represent an HPC workload with increasing importance. In particular the f...
Through the development of large-scale natural language models with writing and dialogue capabilitie...
The teaching and learning group within the School of Computer Science and Mathematics would like to ...
Researchers have been interested in developing AI tools to help students learn various mathematical ...
In today’s world, which is full of innovations in various fields, the role of Information Technologi...
The transformers that drive chatbots and other AI systems constitute large language models (LLMs). T...
In the last decade, the size of deep neural architectures implied in Natural Language Processing (NL...
In his 2016 book, The Fourth Industrial Revolution, Klaus Schwab, the founder of the World Economic ...
In today's world, which is full of innovations in various fields, the role of Information Technologi...
Where Humans Meet Machines: Innovative Solutions for Knotty Natural-Language Problems brings humans ...
The goal of my thesis is to investigate the most influential transformer architectures and to apply ...
Neural machine translation (NMT) is an approach to machine translation (MT) that uses deep learning ...
This report documents the program and the outcomes of Dagstuhl Seminar 22232 "Efficient and Equitabl...
Transformer-based neural models are used in many AI applications. Training these models is expensive...
This report documents the program and the outcomes of Dagstuhl Seminar 22232 “Efficient and Equitabl...
Artificial neural networks represent an HPC workload with increasing importance. In particular the f...
Through the development of large-scale natural language models with writing and dialogue capabilitie...
The teaching and learning group within the School of Computer Science and Mathematics would like to ...
Researchers have been interested in developing AI tools to help students learn various mathematical ...
In today’s world, which is full of innovations in various fields, the role of Information Technologi...
The transformers that drive chatbots and other AI systems constitute large language models (LLMs). T...
In the last decade, the size of deep neural architectures implied in Natural Language Processing (NL...
In his 2016 book, The Fourth Industrial Revolution, Klaus Schwab, the founder of the World Economic ...
In today's world, which is full of innovations in various fields, the role of Information Technologi...
Where Humans Meet Machines: Innovative Solutions for Knotty Natural-Language Problems brings humans ...
The goal of my thesis is to investigate the most influential transformer architectures and to apply ...
Neural machine translation (NMT) is an approach to machine translation (MT) that uses deep learning ...
This report documents the program and the outcomes of Dagstuhl Seminar 22232 "Efficient and Equitabl...
Transformer-based neural models are used in many AI applications. Training these models is expensive...