Thesis (Ph.D.)--University of Washington, 2021Machine learning models, especially deep neural networks, have achieved great success in numerous real-world tasks. As we achieve better performance with larger models, one major challenge emerges that the costs of training machine learning systems become expensive and even prohibitive. Also, the deep learning model works as a block box in many applications with little interpretation of its behaviors. In this dissertation, we investigate two problems: 1) partitioning of training data into diverse and representative blocks for gradient computation to get improved efficiency and performance for machine learning models and 2) decomposition of ReLU deep neural networks as a collection of linear mode...
Recent advances in Artificial Intelligence (AI) are characterized by ever-increasing sizes of datase...
While Deep Neural Networks (DNNs) have recently achieved impressive results on many classification t...
Understanding the computational complexity of training simple neural networks with rectified linear ...
While deep learning continues to advance our technological world, its theoretical underpinnings are ...
Understanding the fundamental principles behind the success of deep neural networks is one of the mo...
We study the inductive bias of two-layer ReLU networks trained by gradient flow. We identify a class...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
In artificial neural networks, learning from data is a computationally demanding task in which a lar...
Thesis (Ph.D.)--University of Washington, 2019Imposing appropriate structure or constraints onto opt...
Rectified linear units (ReLUs) have become the main model for the neural units in current deep learn...
In artificial neural networks, learning from data is a computationally demanding task in which a lar...
The performance decay experienced by deep neural networks (DNNs) when confronted with distributional...
Neural networks (NNs) have seen a surge in popularity due to their unprecedented practical success i...
It is well-known that modern neural networks are vulnerable to adversarial examples. To mitigate thi...
In this thesis, we study model parameterization for deep learning applications. Part of the mathemat...
Recent advances in Artificial Intelligence (AI) are characterized by ever-increasing sizes of datase...
While Deep Neural Networks (DNNs) have recently achieved impressive results on many classification t...
Understanding the computational complexity of training simple neural networks with rectified linear ...
While deep learning continues to advance our technological world, its theoretical underpinnings are ...
Understanding the fundamental principles behind the success of deep neural networks is one of the mo...
We study the inductive bias of two-layer ReLU networks trained by gradient flow. We identify a class...
The success of deep learning has shown impressive empirical breakthroughs, but many theoretical ques...
In artificial neural networks, learning from data is a computationally demanding task in which a lar...
Thesis (Ph.D.)--University of Washington, 2019Imposing appropriate structure or constraints onto opt...
Rectified linear units (ReLUs) have become the main model for the neural units in current deep learn...
In artificial neural networks, learning from data is a computationally demanding task in which a lar...
The performance decay experienced by deep neural networks (DNNs) when confronted with distributional...
Neural networks (NNs) have seen a surge in popularity due to their unprecedented practical success i...
It is well-known that modern neural networks are vulnerable to adversarial examples. To mitigate thi...
In this thesis, we study model parameterization for deep learning applications. Part of the mathemat...
Recent advances in Artificial Intelligence (AI) are characterized by ever-increasing sizes of datase...
While Deep Neural Networks (DNNs) have recently achieved impressive results on many classification t...
Understanding the computational complexity of training simple neural networks with rectified linear ...