Compositional generalization is a basic mechanism in human language learning, which current neural networks struggle with. A recently proposed Disentangled sequence-to-sequence model (Dangle) shows promising generalization capability by learning specialized encodings for each decoding step. We introduce two key modifications to this model which encourage more disentangled representations and improve its compute and memory efficiency, allowing us to tackle compositional generalization in a more realistic setting. Specifically, instead of adaptively re-encoding source keys and values at each time step, we disentangle their representations and only re-encode keys periodically, at some interval. Our new architecture leads to better generalizati...
In the last decade, deep artificial neural networks have achieved astounding performance in many nat...
Funding Information: We thank Yonatan Bisk for his valuable feedback and suggestions on this work. W...
Systematic generalization is the ability to combine known parts into novel meaning; an important asp...
Generic unstructured neural networks have been shown to struggle on out-of-distribution compositiona...
Flexible neural sequence models outperform grammar- and automaton-based counterparts on a variety of...
Humans can systematically generalize to novel compositions of existing concepts. Recent studies argu...
Compositionality---the principle that the meaning of a complex expression is built from the meanings...
Human intelligence exhibits compositional generalization (i.e., the capacity to understand and produ...
While recent work has convincingly showed that sequence-to-sequence models struggle to generalize to...
Neural networks have revolutionized language modeling and excelled in various downstream tasks. Howe...
Seq2seq models have been shown to struggle with compositional generalisation, i.e. generalising to n...
Recombining known primitive concepts into larger novel combinations is a quintessentially human cogn...
When writing programs, people have the ability to tackle a new complex task by decomposing it into s...
The power of human language and thought arises from systematic compositionality—the algebraic abilit...
In tasks like semantic parsing, instruction following, and question answering, standard deep network...
In the last decade, deep artificial neural networks have achieved astounding performance in many nat...
Funding Information: We thank Yonatan Bisk for his valuable feedback and suggestions on this work. W...
Systematic generalization is the ability to combine known parts into novel meaning; an important asp...
Generic unstructured neural networks have been shown to struggle on out-of-distribution compositiona...
Flexible neural sequence models outperform grammar- and automaton-based counterparts on a variety of...
Humans can systematically generalize to novel compositions of existing concepts. Recent studies argu...
Compositionality---the principle that the meaning of a complex expression is built from the meanings...
Human intelligence exhibits compositional generalization (i.e., the capacity to understand and produ...
While recent work has convincingly showed that sequence-to-sequence models struggle to generalize to...
Neural networks have revolutionized language modeling and excelled in various downstream tasks. Howe...
Seq2seq models have been shown to struggle with compositional generalisation, i.e. generalising to n...
Recombining known primitive concepts into larger novel combinations is a quintessentially human cogn...
When writing programs, people have the ability to tackle a new complex task by decomposing it into s...
The power of human language and thought arises from systematic compositionality—the algebraic abilit...
In tasks like semantic parsing, instruction following, and question answering, standard deep network...
In the last decade, deep artificial neural networks have achieved astounding performance in many nat...
Funding Information: We thank Yonatan Bisk for his valuable feedback and suggestions on this work. W...
Systematic generalization is the ability to combine known parts into novel meaning; an important asp...