Deep Neural Networks (DNNs) are increasingly being used in software engineering and code intelligence tasks. These are powerful tools that are capable of learning highly generalizable patterns from large datasets through millions of parameters. At the same time, their large capacity can render them prone to memorizing data points. Recent work suggests that the memorization risk manifests especially strongly when the training dataset is noisy, involving many ambiguous or questionable samples, and memorization is the only recourse. The goal of this paper is to evaluate and compare the extent of memorization and generalization in neural code intelligence models. It aims to provide insights on how memorization may impact the learning behavior o...
“Deep learning” uses Post-Selection—selection of a model after training multiple models using data. ...
Deep learning research has recently witnessed an impressively fast-paced progress in a wide range of...
This electronic version was submitted by the student author. The certified thesis is available in th...
With the prevalence of publicly available source code repositories to train deep neural network mode...
The abundance of publicly available source code repositories, in conjunction with the advances in ne...
The understanding of generalization in machine learning is in a state of flux. This is partly due to...
Over-parameterized deep neural networks (DNNs) with sufficient capacity to memorize random noise can...
Context: With the prevalence of publicly available source code repositories to train deep neural net...
Over-paramaterized neural models have become dominant in Natural Language Processing. Increasing the...
Estimating the Generalization Error (GE) of Deep Neural Networks (DNNs) is an important task that of...
Despite the recent trend of developing and applying neural source code models to software engineerin...
Deep Learning (read neural networks) has emerged as one of the most exciting and powerful tools in t...
State-of-the-art pre-trained language models have been shown to memorise facts and per- form well wi...
This research demonstrates a method of discriminating the numerical relationships of neural network ...
Increasing the size of overparameterized neural networks has been shown to improve their generalizat...
“Deep learning” uses Post-Selection—selection of a model after training multiple models using data. ...
Deep learning research has recently witnessed an impressively fast-paced progress in a wide range of...
This electronic version was submitted by the student author. The certified thesis is available in th...
With the prevalence of publicly available source code repositories to train deep neural network mode...
The abundance of publicly available source code repositories, in conjunction with the advances in ne...
The understanding of generalization in machine learning is in a state of flux. This is partly due to...
Over-parameterized deep neural networks (DNNs) with sufficient capacity to memorize random noise can...
Context: With the prevalence of publicly available source code repositories to train deep neural net...
Over-paramaterized neural models have become dominant in Natural Language Processing. Increasing the...
Estimating the Generalization Error (GE) of Deep Neural Networks (DNNs) is an important task that of...
Despite the recent trend of developing and applying neural source code models to software engineerin...
Deep Learning (read neural networks) has emerged as one of the most exciting and powerful tools in t...
State-of-the-art pre-trained language models have been shown to memorise facts and per- form well wi...
This research demonstrates a method of discriminating the numerical relationships of neural network ...
Increasing the size of overparameterized neural networks has been shown to improve their generalizat...
“Deep learning” uses Post-Selection—selection of a model after training multiple models using data. ...
Deep learning research has recently witnessed an impressively fast-paced progress in a wide range of...
This electronic version was submitted by the student author. The certified thesis is available in th...