<p>The second column gives the performance of an Elastic Net model under cross-validation on the training set. The third column gives the performance on the held-out test set, with 95% confidence intervals determined using the bias-corrected and accelerated bootstrap. The nearly identical overlap of the confidence intervals indicates that the classifiers built from each of the two learned feature layers and the expert-engineered feature set were equally useful in the supervised learning task. Likewise, the 0.04 difference in performance between the baseline model and the other three is both statistically significant and a respectable improvement as supervised models go. AUC: Area under the Receiver Operating Characteristic curve. CI: 95% Co...
A. Subtask-averaged learning curves for humans and strong baseline models. The y-axis is the percent...
(A) shows the AUC (95% confidence interval) of the p50 model tested on the same subset from which it...
<p>Performance on the benchmark training dataset was evaluated based on AUC, MCC, Accuracy, Specific...
A, B, C, and D: Each colored line indicates a result from a single random training-test set split, a...
<p>(A) Cross-validation (CV) performance of models trained on all available native IRES sequences sh...
<p>We trained a classifier to predict phase III clinical trial outcomes, using 5-fold cross-validati...
<p><b>(A)</b> Receiver operating characteristic (ROC) curve. The solid black line indicates the medi...
The training scores (R2) and cross validation (CV) scores (also R2) are shown. Below 800 training ex...
<p>A: The results of 2-fold cross validations are shown for each regression method in this in this f...
The task performance (task AUC and task accuracy) shows how well classifiers are able to distinguish...
<p>(The feature sets are labeled by their size, here 6 features, and then enumerated from 0 to N–1, ...
(a) and (b) show the box plot of the five-class classification accuracy with RF and SVM, respectivel...
We fit 500 models with 500 subsets of 25 randomly selected features from the uncorrelated 4800 featu...
Results are shown for the top three cross-validated models plus the cross-validated performance of t...
Cross-validated AUC point estimates and 95% confidence intervals are shown for A) models trained on ...
A. Subtask-averaged learning curves for humans and strong baseline models. The y-axis is the percent...
(A) shows the AUC (95% confidence interval) of the p50 model tested on the same subset from which it...
<p>Performance on the benchmark training dataset was evaluated based on AUC, MCC, Accuracy, Specific...
A, B, C, and D: Each colored line indicates a result from a single random training-test set split, a...
<p>(A) Cross-validation (CV) performance of models trained on all available native IRES sequences sh...
<p>We trained a classifier to predict phase III clinical trial outcomes, using 5-fold cross-validati...
<p><b>(A)</b> Receiver operating characteristic (ROC) curve. The solid black line indicates the medi...
The training scores (R2) and cross validation (CV) scores (also R2) are shown. Below 800 training ex...
<p>A: The results of 2-fold cross validations are shown for each regression method in this in this f...
The task performance (task AUC and task accuracy) shows how well classifiers are able to distinguish...
<p>(The feature sets are labeled by their size, here 6 features, and then enumerated from 0 to N–1, ...
(a) and (b) show the box plot of the five-class classification accuracy with RF and SVM, respectivel...
We fit 500 models with 500 subsets of 25 randomly selected features from the uncorrelated 4800 featu...
Results are shown for the top three cross-validated models plus the cross-validated performance of t...
Cross-validated AUC point estimates and 95% confidence intervals are shown for A) models trained on ...
A. Subtask-averaged learning curves for humans and strong baseline models. The y-axis is the percent...
(A) shows the AUC (95% confidence interval) of the p50 model tested on the same subset from which it...
<p>Performance on the benchmark training dataset was evaluated based on AUC, MCC, Accuracy, Specific...