High accuracy forecasts are essential to financial risk management, where machine learning algorithms are frequently employed. We derive a new theoretical bound on the sample complexity for Probably Approximately Correct (PAC) learning in the presence of noise, and does not require specification of the hypothesis set |H|. We demonstrate that for realistic financial applications where |H| is typically infinite. This is contrary to prior theoretical conclusions. We further show that noise, which is a non-trivial component of big data, has a dominating impact on the data size required for PAC learning. Consequently, contrary to current big data trends, we argue that high quality data is more important than large volumes of data. This paper add...
The Vapnik-Chervonenkis (VC) dimension is a combinatorial measure of a certain class of machine lear...
We information-theoretically reformulate two measures of capacity from statistical learning theory: ...
htmlabstractWe present a novel notion of complexity that interpolates between and generalizes some c...
We discuss more realistic models of computational learning. We extend the existing literature on the...
We narrow the width of the confidence interval introduced by Vapnik and Chervonenkis for the risk fu...
International audienceMachine learning algorithms and big data are transforming all industries inclu...
AbstractWhat we learned from the global financial crisis is that to get information about the underl...
Price movements in financial markets are well known to be very noisy. As a result, even if there are...
AbstractThe PAC model of learning and its extension to real valued function classes provides a well-...
Risk bounds, which are also called generalisation bounds in the statistical learning literature, are...
We report new results about the impact of noise on information processing with application to financ...
AbstractThis paper presents a general information-theoretic approach for obtaining lower bounds on t...
This thesis focuses on sources of error in modern Bayesian analysis and machine learning in the ``bi...
We study the performance -- and specifically the rate at which the error probability converges to ze...
This thesis addresses three challenge of machine learning: high-dimensional data, label noise and li...
The Vapnik-Chervonenkis (VC) dimension is a combinatorial measure of a certain class of machine lear...
We information-theoretically reformulate two measures of capacity from statistical learning theory: ...
htmlabstractWe present a novel notion of complexity that interpolates between and generalizes some c...
We discuss more realistic models of computational learning. We extend the existing literature on the...
We narrow the width of the confidence interval introduced by Vapnik and Chervonenkis for the risk fu...
International audienceMachine learning algorithms and big data are transforming all industries inclu...
AbstractWhat we learned from the global financial crisis is that to get information about the underl...
Price movements in financial markets are well known to be very noisy. As a result, even if there are...
AbstractThe PAC model of learning and its extension to real valued function classes provides a well-...
Risk bounds, which are also called generalisation bounds in the statistical learning literature, are...
We report new results about the impact of noise on information processing with application to financ...
AbstractThis paper presents a general information-theoretic approach for obtaining lower bounds on t...
This thesis focuses on sources of error in modern Bayesian analysis and machine learning in the ``bi...
We study the performance -- and specifically the rate at which the error probability converges to ze...
This thesis addresses three challenge of machine learning: high-dimensional data, label noise and li...
The Vapnik-Chervonenkis (VC) dimension is a combinatorial measure of a certain class of machine lear...
We information-theoretically reformulate two measures of capacity from statistical learning theory: ...
htmlabstractWe present a novel notion of complexity that interpolates between and generalizes some c...