In this study, we propose a new approach for novelty detection that uses kernel dependence techniques for characterizing the statistical dependencies of random variables (RV) and use this characterization as a basis for making inference. Considering the statistical dependencies of the RVs in multivariate problems is an important challenge in novelty detection. Ignoring these dependencies, when they are strong, may result in inaccurate inference, usually in the form of high false positive rates. Previously studied methods, such as graphical models or conditional classifiers, mainly use density estimation techniques as their main learning element to characterize the dependencies of the relevant RVs. Therefore, they suffer from the curse of di...