International audienceAlthough deep pre-trained language models have shown promising benefit in a large set of industrial scenarios, including Click-Through-Rate (CTR) prediction, how to integrate pre-trained language models that handle only textual signals into a prediction pipeline with non-textual features is challenging. Up to now, two directions have been explored to integrate multimodal inputs in fine-tuning of pre-trained language models. One consists of fusing the outcome of language models and non-textual features through an aggregation layer, resulting into ensemble framework, where the cross-information between textual and nontextual inputs are learned only in the aggregation layer. The second one consists of splitting and transf...
International audienceIn this paper, we address a relatively new task: prediction of ASR performance...
Current language models have been criticised for learning language from text alone without connectio...
Despite advances in deep learning and knowledge graphs (KGs), using language models for natural lang...
International audienceIn sponsored search engines, pre-trained language models have shown promising ...
Traditional click-through rate (CTR) prediction models convert the tabular data into one-hot vectors...
Self-supervised pre-training of language models usually consists in predicting probability distribut...
The click-through rate (CTR) prediction task is to predict whether a user will click on the recommen...
When pre-trained on large unsupervised textual corpora, language models are able to store and retri...
CTR prediction has been widely used in the real world. Many methods model feature interaction to imp...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
International audienceWe probe pre-trained transformer language models for bridging inference. We fi...
Convolutional Neural Networks (CNNs) have shown to yield very strong results in several Computer Vis...
International audienceThis paper addresses a relatively new task: prediction of ASR performance on u...
International audienceIn this paper we describe our contribution to the CMCL 2021 Shared Task, which...
Language modeling is a very broad field and has been used for various purposes for a long period of ...
International audienceIn this paper, we address a relatively new task: prediction of ASR performance...
Current language models have been criticised for learning language from text alone without connectio...
Despite advances in deep learning and knowledge graphs (KGs), using language models for natural lang...
International audienceIn sponsored search engines, pre-trained language models have shown promising ...
Traditional click-through rate (CTR) prediction models convert the tabular data into one-hot vectors...
Self-supervised pre-training of language models usually consists in predicting probability distribut...
The click-through rate (CTR) prediction task is to predict whether a user will click on the recommen...
When pre-trained on large unsupervised textual corpora, language models are able to store and retri...
CTR prediction has been widely used in the real world. Many methods model feature interaction to imp...
Natural language processing (NLP) techniques had significantly improved by introducing pre-trained l...
International audienceWe probe pre-trained transformer language models for bridging inference. We fi...
Convolutional Neural Networks (CNNs) have shown to yield very strong results in several Computer Vis...
International audienceThis paper addresses a relatively new task: prediction of ASR performance on u...
International audienceIn this paper we describe our contribution to the CMCL 2021 Shared Task, which...
Language modeling is a very broad field and has been used for various purposes for a long period of ...
International audienceIn this paper, we address a relatively new task: prediction of ASR performance...
Current language models have been criticised for learning language from text alone without connectio...
Despite advances in deep learning and knowledge graphs (KGs), using language models for natural lang...