Transformers are the current state-of-the-art of natural language processing in many domains and are using traction within software engineering research as well. Such models are pre-trained on large amounts of data, usually from the general domain. However, we only have a limited understanding regarding the validity of transformers within the software engineering domain, i.e., how good such models are at understanding words and sentences within a software engineering context and how this improves the state-of-the-art. Within this article, we shed light on this complex, but crucial issue. We compare BERT transformer models trained with software engineering data with transformers based on general domain data in multiple dimensions: their voca...
This document aims to be a self-contained, mathematically precise overview of transformer architectu...
Natural language generation (nlg) systems are computer software systems that pro-duce texts in Engli...
This paper reports on the evaluation of Deep Learning (DL) transformer architecture models for Named...
The Bidirectional Encoder Representations from Transformers (BERT) is currently one of the most impo...
In the last decade, the size of deep neural architectures implied in Natural Language Processing (NL...
The software development process produces vast amounts of textual data expressed in natural language...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
In today’s world, which is full of innovations in various fields, the role of Information Technologi...
Pre-trained transformers have rapidly become very popular in the Natural Language Processing (NLP) c...
Large Transformer models achieved the state-of-the-art status for Natural Language Understanding tas...
Software development is a complex activity that requires, in addition to professional knowledge and ...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Transformer-based masked language models trained on general corpora, such as BERT and RoBERTa, have ...
In today's world, which is full of innovations in various fields, the role of Information Technologi...
Artificial intelligence (AI) for software engineering (SE) tasks has recently achieved promising per...
This document aims to be a self-contained, mathematically precise overview of transformer architectu...
Natural language generation (nlg) systems are computer software systems that pro-duce texts in Engli...
This paper reports on the evaluation of Deep Learning (DL) transformer architecture models for Named...
The Bidirectional Encoder Representations from Transformers (BERT) is currently one of the most impo...
In the last decade, the size of deep neural architectures implied in Natural Language Processing (NL...
The software development process produces vast amounts of textual data expressed in natural language...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
In today’s world, which is full of innovations in various fields, the role of Information Technologi...
Pre-trained transformers have rapidly become very popular in the Natural Language Processing (NLP) c...
Large Transformer models achieved the state-of-the-art status for Natural Language Understanding tas...
Software development is a complex activity that requires, in addition to professional knowledge and ...
Recently, the development of pre-trained language models has brought natural language processing (NL...
Transformer-based masked language models trained on general corpora, such as BERT and RoBERTa, have ...
In today's world, which is full of innovations in various fields, the role of Information Technologi...
Artificial intelligence (AI) for software engineering (SE) tasks has recently achieved promising per...
This document aims to be a self-contained, mathematically precise overview of transformer architectu...
Natural language generation (nlg) systems are computer software systems that pro-duce texts in Engli...
This paper reports on the evaluation of Deep Learning (DL) transformer architecture models for Named...