Machine-learning models can reach very high performance with supervised training, where they learn from labeled data. However, supervised training requires annotating data with desired output labels, which can be a difficult and time-consuming task. Meanwhile, advancements in deep learning models and technology have made it possible to train very large models, which was not feasible a few years ago. Although training such big models requires a substantial amount of supervised data, models can overcome this limitation by first learning from un-labeled data. Pre-trained language models enable us to achieve state-of-the-art performance from large-scale models with limited supervised data. During the pre-training stage, models are exposed to un...
Large Language Models (LLM) are a new class of computation engines, "programmed" via prompt engineer...
The way software developers edit code day-to-day tends to be repetitive, often using existing code e...
This dissertation addresses two significant challenges of large language models (LLMs): robustness a...
Pretraining deep neural networks to perform language modeling - that is, to reconstruct missing word...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Thesis (Ph.D.)--University of Washington, 2022A robust language processing machine should be able to...
Improving developer productivity is an important, but very difficult task, that researchers from bot...
© 2018 Association for Computing Machinery. Code summarization provides a high level natural languag...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
Pre-training language models (LMs) on large-scale unlabeled text data makes the model much easier to...
Machine Learning methods, especially Deep Learning, had an enormous breakthrough in Natural Language...
Predictive modeling using machine learning is an effective method for building compiler heuristics, ...
Large Language Models (LLMs) play an ever-increasing role in the field of Artificial Intelligence (A...
Within a situation where Semi-Supervised Learning (SSL) is available to exploit unlabeled data, this...
State-of-the-art pre-trained language models have been shown to memorise facts and per- form well wi...
Large Language Models (LLM) are a new class of computation engines, "programmed" via prompt engineer...
The way software developers edit code day-to-day tends to be repetitive, often using existing code e...
This dissertation addresses two significant challenges of large language models (LLMs): robustness a...
Pretraining deep neural networks to perform language modeling - that is, to reconstruct missing word...
This paper studies the use of language models as a source of synthetic unlabeled text for NLP. We fo...
Thesis (Ph.D.)--University of Washington, 2022A robust language processing machine should be able to...
Improving developer productivity is an important, but very difficult task, that researchers from bot...
© 2018 Association for Computing Machinery. Code summarization provides a high level natural languag...
International audienceLarge language models have recently been shown to attain reasonable zero-shot ...
Pre-training language models (LMs) on large-scale unlabeled text data makes the model much easier to...
Machine Learning methods, especially Deep Learning, had an enormous breakthrough in Natural Language...
Predictive modeling using machine learning is an effective method for building compiler heuristics, ...
Large Language Models (LLMs) play an ever-increasing role in the field of Artificial Intelligence (A...
Within a situation where Semi-Supervised Learning (SSL) is available to exploit unlabeled data, this...
State-of-the-art pre-trained language models have been shown to memorise facts and per- form well wi...
Large Language Models (LLM) are a new class of computation engines, "programmed" via prompt engineer...
The way software developers edit code day-to-day tends to be repetitive, often using existing code e...
This dissertation addresses two significant challenges of large language models (LLMs): robustness a...