Understanding the black-box prediction for neural networks is challenging. To achieve this, early studies have designed influence function (IF) to measure the effect of removing a single training point on neural networks. However, the classic implicit Hessian-vector product (IHVP) method for calculating IF is fragile, and theoretical analysis of IF in the context of neural networks is still lacking. To this end, we utilize the neural tangent kernel (NTK) theory to calculate IF for the neural network trained with regularized mean-square loss, and prove that the approximation error can be arbitrarily small when the width is sufficiently large for two-layer ReLU networks. We analyze the error bound for the classic IHVP method in the over-param...
Polynomial neural networks (NNs-Hp) have recently demonstrated high expressivity and efficiency acro...
Recently, several studies have proven the global convergence and generalization abilities of the gra...
Neural networks are more expressive when they have multiple layers. In turn, conventional training m...
International audienceState-of-the-art neural networks are heavily over-parameterized, making the op...
Neural tangent kernel (NTK) is a powerful tool to analyze training dynamics of neural networks and t...
Parameter estimation in empirical fields is usually undertaken using parametric models, and such mod...
Small generalization errors of over-parameterized neural networks (NNs) can be partially explained b...
In the recent years, Deep Neural Networks (DNNs) have managed to succeed at tasks that previously ap...
Value approximation using deep neural networks is at the heart of off-policy deep reinforcement lear...
A trained neural network can be interpreted as a structural causal model (SCM) that provides the eff...
International audienceWe investigate the robustness of feed-forward neural networks when input data ...
Recently, neural tangent kernel (NTK) has been used to explain the dynamics of learning parameters o...
We establish PAC learnability of influence functions for three common influence models, namely, the ...
We describe the notion of "equivalent kernels " and suggest that this provides a framework...
It took until the last decade to finally see a machine match human performance on essentially any ta...
Polynomial neural networks (NNs-Hp) have recently demonstrated high expressivity and efficiency acro...
Recently, several studies have proven the global convergence and generalization abilities of the gra...
Neural networks are more expressive when they have multiple layers. In turn, conventional training m...
International audienceState-of-the-art neural networks are heavily over-parameterized, making the op...
Neural tangent kernel (NTK) is a powerful tool to analyze training dynamics of neural networks and t...
Parameter estimation in empirical fields is usually undertaken using parametric models, and such mod...
Small generalization errors of over-parameterized neural networks (NNs) can be partially explained b...
In the recent years, Deep Neural Networks (DNNs) have managed to succeed at tasks that previously ap...
Value approximation using deep neural networks is at the heart of off-policy deep reinforcement lear...
A trained neural network can be interpreted as a structural causal model (SCM) that provides the eff...
International audienceWe investigate the robustness of feed-forward neural networks when input data ...
Recently, neural tangent kernel (NTK) has been used to explain the dynamics of learning parameters o...
We establish PAC learnability of influence functions for three common influence models, namely, the ...
We describe the notion of "equivalent kernels " and suggest that this provides a framework...
It took until the last decade to finally see a machine match human performance on essentially any ta...
Polynomial neural networks (NNs-Hp) have recently demonstrated high expressivity and efficiency acro...
Recently, several studies have proven the global convergence and generalization abilities of the gra...
Neural networks are more expressive when they have multiple layers. In turn, conventional training m...