Logo Violin Plot Heatmaps Regression plots: True vs. Predicted Correlation comparison Correlation comparison per drug Correlation comparison per cell line Tables
Logo

Results for ldo_run: LDO



This diagram displays the mean rank of each model over all cross-validation splits: Within each CV split, the models are ranked according to their MSE. We calculate whether a model is significantly better than another one using the Friedman test and the post-hoc Conover test. The Friedman test shows whether there are overall differences between the models. After a significantFriedman test, the pairwise Conover test is performed to identify which models are significantly outperforming others. One line indicates which models are not significantly different from each other. The p-values are shown below. This can only be rendered if at least 3 models were run.

Results of Post-Hoc Conover Test


DIPK ElasticNet GradientBoosting MultiOmicsNeuralNetwork MultiOmicsRandomForest NaiveMeanEffectsPredictor NaivePredictor RandomForest SimpleNeuralNetwork
DIPK 1.0000 0.9345 0.3675 0.8695 1.0000 0.8053 0.1913 0.1422 0.6814
ElasticNet 0.9345 1.0000 0.3258 0.9345 0.9345 0.7425 0.2202 0.1653 0.6223
GradientBoosting 0.3675 0.3258 1.0000 0.2874 0.3675 0.5116 0.0291 0.0194 0.6223
MultiOmicsNeuralNetwork 0.8695 0.9345 0.2874 1.0000 0.8695 0.6814 0.2522 0.1913 0.5657
MultiOmicsRandomForest 1.0000 0.9345 0.3675 0.8695 1.0000 0.8053 0.1913 0.1422 0.6814
NaiveMeanEffectsPredictor 0.8053 0.7425 0.5116 0.6814 0.8053 1.0000 0.1216 0.0877 0.8695
NaivePredictor 0.1913 0.2202 0.0291 0.2522 0.1913 0.1216 1.0000 0.8695 0.0877
RandomForest 0.1422 0.1653 0.0194 0.1913 0.1422 0.0877 0.8695 1.0000 0.0619
SimpleNeuralNetwork 0.6814 0.6223 0.6223 0.5657 0.6814 0.8695 0.0877 0.0619 1.0000

Violin Plots of Performance Measures over CV runs

Violin plots comparing all models

To focus on a specific metric, choose it in the dropdown menu in the top right corner.You can investigate the distribution of the performance measures by hovering over the plot. To select/exclude specific algorithms, (double-)click them in the legend.

Violin plots comparing all models with normalized metrics

Before calculating the evaluation metrics, all values were normalized by the predictions of the NaiveMeanEffectsPredictor. Since this only influences the R^2 and the correlation metrics, the error metrics are not shown.

Violin plots comparing performance measures for tests within each model

Heatmap Plots of Performance Measures over CV runs

Heatmap plots comparing all models

Unnormalized metrics collapsed over all CV runs with mean and standard deviation. The strictly standardized mean difference is a measure of effect size which is calculated pairwise. For two models, it is calculated as [mean1 - mean2] / [sqrt(var1 + var2)] for a specific measure. The larger the absolute SSMD, the stronger the effect (a strong effect could, is e.g., a |SSMD| > 2 ).

Heatmap plots comparing all models with normalized metrics

Before calculating the evaluation metrics, all values were normalized by the predictions of the NaiveMeanEffectsPredictor. Since this only influences the R^2 and the correlation metrics, the error metrics are not shown.

Heatmap plots comparing performance measures for tests within each model

Regression plots

Comparison of normalized R^2 values

R^2 values can be compared here between models, either per cell line or per drug. This can either show if a model has consistently higher or lower R^2 values than another model or identify cell lines/drugs for which models agree or disagree. The x-axis is the first dropdown menu, the y-axis is the second dropdown menu.

Drug_name-wise comparison

Comparisons per model