Abstract. Universal Consistency, the convergence to the minimum possible er-ror rate in learning through genetic programming (GP), and Code bloat, the ex-cessive increase of code size, are important issues in GP. This paper proposes a theoretical analysis of universal consistency and code bloat in the framework of symbolic regression in GP, from the viewpoint of Statistical Learning The-ory, a well grounded mathematical toolbox for Machine Learning. Two kinds of bloat must be distinguished in that context, depending whether the target function has finite description length or not. Then, the Vapnik-Chervonenkis dimension of programs is computed, and we prove that a parsimonious fitness ensures Uni-versal Consistency (i.e. the fact that the s...