Deep neural networks have achieved significant success in a number of challenging engineering problems. There is consensus in the community that some form of smoothing of the loss function is needed, and there have been hundreds of papers and many conferences in the past three years on this topic. However, so far there has been little analysis by mathematicians. The fundamental tool in training deep neural networks is Stochastic Gradient Descent (SGD) applied to the ``loss'' function, $f(x)$, which is high dimensional and nonconvex. \begin{equation}\label{SGDintro}\tag{SDG} dx_t = -\nabla f(x_t) dt + dW_t \end{equation} There is a consensus in the field that some for of regularization of the loss function is needed, but so far there...
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and i...
In this thesis, we study model parameterization for deep learning applications. Part of the mathemat...
In Theory IIb we characterize with a mix of theory and experiments the optimization of deep convolut...
Deep neural networks have achieved significant success in a number of challenging engineering proble...
In the past decade, neural networks have demonstrated impressive performance in supervised learning....
The classical statistical learning theory implies that fitting too many parameters leads to overfitt...
The deep learning optimization community has observed how the neural networks generalization ability...
Training deep neural networks is inherently subject to the predefined and fixed loss functions durin...
Training deep neural networks is inherently subject to the predefined and fixed loss functions durin...
Recent works have shown that deep neural networks can be employed to solve partial differential equa...
This thesis characterizes the training process of deep neural networks. We are driven by two apparen...
Recent works have shown that deep neural networks can be employed to solve partial differential equa...
Recent works have shown that deep neural networks can be employed to solve partial differential equa...
Recent works have shown that deep neural networks can be employed to solve partial differential equa...
Loss landscape is a useful tool to characterize and compare neural network models. The main challeng...
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and i...
In this thesis, we study model parameterization for deep learning applications. Part of the mathemat...
In Theory IIb we characterize with a mix of theory and experiments the optimization of deep convolut...
Deep neural networks have achieved significant success in a number of challenging engineering proble...
In the past decade, neural networks have demonstrated impressive performance in supervised learning....
The classical statistical learning theory implies that fitting too many parameters leads to overfitt...
The deep learning optimization community has observed how the neural networks generalization ability...
Training deep neural networks is inherently subject to the predefined and fixed loss functions durin...
Training deep neural networks is inherently subject to the predefined and fixed loss functions durin...
Recent works have shown that deep neural networks can be employed to solve partial differential equa...
This thesis characterizes the training process of deep neural networks. We are driven by two apparen...
Recent works have shown that deep neural networks can be employed to solve partial differential equa...
Recent works have shown that deep neural networks can be employed to solve partial differential equa...
Recent works have shown that deep neural networks can be employed to solve partial differential equa...
Loss landscape is a useful tool to characterize and compare neural network models. The main challeng...
Physics-informed neural networks (PINNs) have effectively been demonstrated in solving forward and i...
In this thesis, we study model parameterization for deep learning applications. Part of the mathemat...
In Theory IIb we characterize with a mix of theory and experiments the optimization of deep convolut...