Permutations and matchings are core building blocks in a variety of latent variable models, as they allow us to align, canonicalize, and sort data. Learning in such models is difficult, however, because exact marginalization over these combinatorial objects is intractable. In response, this paper introduces a collection of new methods for end-to-end learning in such models that approximate discrete maximum-weight matching using the continuous Sinkhorn operator. Sinkhorn iteration is attractive because it functions as a simple, easy-to-implement analog of the softmax operator. With this, we can define the Gumbel-Sinkhorn method, an extension of the Gumbel-Softmax method (Jang et al. 2016, Maddison2016 et al. 2016) to distributions over late...
We present a principled approach to uncover the structure of visual data by solving a deep learning ...
With a focus on designing flexible, tractable, and adaptive methodology for some canonical machine l...
In this dissertation, we explore two fundamental sets of inference problems arising in machine learn...
Abstract: It is of increasing importance to develop learning meth-ods for ranking. In contrast to ma...
Numerous models for supervised and reinforcement learning benefit from combinations of discrete and ...
In this thesis we describe two separate works: higher order permutation equivariant layers for neura...
Entropy regularized optimal transport (EOT) distance and its symmetric normalization, known as the S...
Entropy regularized optimal transport (EOT) distance and its symmetric normalization, known as the S...
The overaching goal in this thesis is to develop the representational frameworks, the inference algo...
Graph neural networks (GNNs) have recently emerged as a dominant paradigm for machine learning with ...
Distributional reinforcement learning~(RL) is a class of state-of-the-art algorithms that estimate t...
We formulate the weighted b-matching objective function as a probability distribution function and p...
Modern machine learning relies on algorithms that fit expressive latent models to large datasets. Wh...
ListNet is a well-known listwise learning to rank model and has gained much atten-tion in recent yea...
Existing techniques for improving scalability of weight learning in Markov Logic Networks (MLNs) are...
We present a principled approach to uncover the structure of visual data by solving a deep learning ...
With a focus on designing flexible, tractable, and adaptive methodology for some canonical machine l...
In this dissertation, we explore two fundamental sets of inference problems arising in machine learn...
Abstract: It is of increasing importance to develop learning meth-ods for ranking. In contrast to ma...
Numerous models for supervised and reinforcement learning benefit from combinations of discrete and ...
In this thesis we describe two separate works: higher order permutation equivariant layers for neura...
Entropy regularized optimal transport (EOT) distance and its symmetric normalization, known as the S...
Entropy regularized optimal transport (EOT) distance and its symmetric normalization, known as the S...
The overaching goal in this thesis is to develop the representational frameworks, the inference algo...
Graph neural networks (GNNs) have recently emerged as a dominant paradigm for machine learning with ...
Distributional reinforcement learning~(RL) is a class of state-of-the-art algorithms that estimate t...
We formulate the weighted b-matching objective function as a probability distribution function and p...
Modern machine learning relies on algorithms that fit expressive latent models to large datasets. Wh...
ListNet is a well-known listwise learning to rank model and has gained much atten-tion in recent yea...
Existing techniques for improving scalability of weight learning in Markov Logic Networks (MLNs) are...
We present a principled approach to uncover the structure of visual data by solving a deep learning ...
With a focus on designing flexible, tractable, and adaptive methodology for some canonical machine l...
In this dissertation, we explore two fundamental sets of inference problems arising in machine learn...