| Literature DB >> 29536822 |
Diego Fioravanti1,2, Ylenia Giarratano3, Valerio Maggio1, Claudio Agostinelli4, Marco Chierici1, Giuseppe Jurman5, Cesare Furlanello1.
Abstract
BACKGROUND: Convolutional Neural Networks can be effectively used only when data are endowed with an intrinsic concept of neighbourhood in the input space, as is the case of pixels in images. We introduce here Ph-CNN, a novel deep learning architecture for the classification of metagenomics data based on the Convolutional Neural Networks, with the patristic distance defined on the phylogenetic tree being used as the proximity measure. The patristic distance between variables is used together with a sparsified version of MultiDimensional Scaling to embed the phylogenetic tree in a Euclidean space.Entities:
Keywords: Convolutional neural networks; Deep learning; Metagenomics; Phylogenetic trees
Mesh:
Year: 2018 PMID: 29536822 PMCID: PMC5850953 DOI: 10.1186/s12859-018-2033-5
Source DB: PubMed Journal: BMC Bioinformatics ISSN: 1471-2105 Impact factor: 3.169
Fig. 1Patristic distance on a tree
Fig. 2The structure of Ph-CNN. In this configuration, Ph-CNN is composed by two PhyloConv layers followed by a Fully Connected layer before decision
Fig. 3Data Analysis Protocol for the experimental framework
Patient stratification in the IBD dataset
| HS | IBD patients | |||||
|---|---|---|---|---|---|---|
| CDf | CDr | UCf | UCr | |||
| iCDf | cCDf | iCDr | cCDr | |||
| 38 | 44 | 16 | 59 | 18 | 41 | 44 |
| 14.6% | 16.9% | 6.1% | 22.7% | 6.9% | 15.8% | 16.9% |
Fig. 4Classification tasks on IDB dataset. The six learning tasks discriminating HS versus different stages of IBD patients
Fig. 5The phylogenetic tree for the IDB dataset
Fig. 6Principal component analysis for the 4 synthetic datasets D0,D1,D2,D3, with same sample sizes as in the IBD dataset. Larger values of α correspond to more separate classes HR and CDf
Dataset D: classification performance of Ph-CNN compared to other classifiers on Healthy vs. UCf classification task
| UCf | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 63 | 0.794 | 0.785 | 0.803 | 0.799 | 0.793 | 0.803 |
| 125 | 0.852 | 0.845 | 0.860 | 0.861 | 0.857 | 0.865 |
| 188 | 0.920 | 0.916 | 0.925 | 0.924 | 0.921 | 0.926 |
| 250 | 0.940 | 0.937 | 0.944 | 0.943 | 0.941 | 0.945 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 63 | 0.701 | 0.692 | 0.721 | 0.729 | 0.723 | 0.736 |
| 125 | 0.838 | 0.834 | 0.842 | 0.843 | 0.837 | 0.849 |
| 188 | 0.865 | 0.861 | 0.869 | 0.902 | 0.899 | 0.906 |
| 250 | 0.898 | 0.894 | 0.901 | 0.903 | 0.900 | 0.907 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D: classification performance of Ph-CNN compared to other classifiers on Healthy vs. UCr classification task
| UCr | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 60 | 0.861 | 0.855 | 0.867 | 0.811 | 0.807 | 0.815 |
| 119 | 0.893 | 0.888 | 0.899 | 0.866 | 0.862 | 0.870 |
| 178 | 0.906 | 0.900 | 0.911 | 0.892 | 0.888 | 0.895 |
| 237 | 0.920 | 0.916 | 0.924 | 0.917 | 0.914 | 0.920 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 60 | 0.873 | 0.869 | 0.443 | 0.797 | 0.792 | 0.801 |
| 119 | 0.877 | 0.873 | 0.877 | 0.799 | 0.794 | 0.803 |
| 178 | 0.859 | 0.855 | 0.880 | 0.791 | 0.787 | 0.794 |
| 237 | 0.849 | 0.844 | 0.854 | 0.790 | 0.786 | 0.795 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D: classification performance of Ph-CNN compared to other classifiers on Healthy vs. CDf classification task
| CDf | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.785 | 0.775 | 0.795 | 0.781 | 0.776 | 0.785 |
| 130 | 0.832 | 0.825 | 0.840 | 0.833 | 0.829 | 0.838 |
| 195 | 0.896 | 0.891 | 0.901 | 0.910 | 0.907 | 0.912 |
| 259 | 0.927 | 0.924 | 0.930 | 0.920 | 0.918 | 0.923 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.604 | 0.593 | 0.614 | 0.764 | 0.760 | 0.769 |
| 130 | 0.821 | 0.817 | 0.825 | 0.805 | 0.800 | 0.810 |
| 195 | 0.830 | 0.825 | 0.836 | 0.863 | 0.860 | 0.867 |
| 259 | 0.858 | 0.854 | 0.862 | 0.880 | 0.877 | 0.883 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D: classification performance of Ph-CNN compared to other classifiers on Healthy vs. CDr classification task
| CDr | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.714 | 0.705 | 0.723 | 0.740 | 0.734 | 0.746 |
| 129 | 0.799 | 0.793 | 0.806 | 0.802 | 0.798 | 0.808 |
| 193 | 0.850 | 0.844 | 0.856 | 0.860 | 0.857 | 0.864 |
| 257 | 0.890 | 0.884 | 0.895 | 0.880 | 0.877 | 0.882 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.498 | 0.473 | 0.521 | 0.688 | 0.682 | 0.695 |
| 129 | 0.783 | 0.778 | 0.788 | 0.744 | 0.740 | 0.784 |
| 193 | 0.766 | 0.759 | 0.773 | 0.762 | 0.756 | 0.767 |
| 257 | 0.788 | 0.782 | 0.794 | 0.765 | 0.761 | 0.771 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D: classification performance of Ph-CNN compared to other classifiers on Healthy vs. iCDf classification task
| iCDf | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 62 | 0.781 | 0.772 | 0.790 | 0.804 | 0.799 | 0.808 |
| 124 | 0.863 | 0.854 | 0.871 | 0.861 | 0.858 | 0.865 |
| 186 | 0.922 | 0.918 | 0.926 | 0.921 | 0.919 | 0.924 |
| 247 | 0.944 | 0.941 | 0.947 | 0.941 | 0.939 | 0.942 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 62 | 0.845 | 0.840 | 0.849 | 0.748 | 0.743 | 0.753 |
| 124 | 0.889 | 0.886 | 0.893 | 0.808 | 0.803 | 0.814 |
| 186 | 0.879 | 0.875 | 0.883 | 0.880 | 0.877 | 0.883 |
| 247 | 0.901 | 0.899 | 0.904 | 0.890 | 0.887 | 0.893 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D: classification performance of Ph-CNN compared to other classifiers on Healthy vs. iCDr classification task
| iCDr | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.753 | 0.744 | 0.763 | 0.773 | 0.769 | 0.779 |
| 129 | 0.830 | 0.823 | 0.837 | 0.834 | 0.830 | 0.837 |
| 193 | 0.884 | 0.878 | 0.889 | 0.893 | 0.891 | 0.896 |
| 257 | 0.910 | 0.905 | 0.915 | 0.907 | 0.904 | 0.909 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 63 | 0.807 | 0.802 | 0.812 | 0.724 | 0.719 | 0.729 |
| 125 | 0.822 | 0.816 | 0.827 | 0.794 | 0.788 | 0.800 |
| 188 | 0.831 | 0.827 | 0.835 | 0.812 | 0.807 | 0.818 |
| 250 | 0.837 | 0.831 | 0.842 | 0.820 | 0.816 | 0.825 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D: classification performance of Ph-CNN compared to other classifiers on the external validation dataset
| Task | Ph-CNN | LSVM | MLPNN | RF |
|---|---|---|---|---|
| UCf | 0.946 | 0.934 | 0.898 | 0.869 |
| UCr | 0.897 | 0.904 | 0.897 | 0.756 |
| CDf | 0.926 | 0.935 | 0.884 | 0.859 |
| CDr | 0.888 | 0.888 | 0.821 | 0.722 |
| iCDf | 0.931 | 0.943 | 0.905 | 0.863 |
| iCDr | 0.901 | 0.910 | 0.846 | 0.778 |
Dataset D on IBD: classification performance of Ph-CNN compared to other classifiers on Healthy vs. UCf classification task
| UCf | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 63 | 0.659 | 0.604 | 0.709 | 0.510 | 0.449 | 0.573 |
| 125 | 0.668 | 0.595 | 0.734 | 0.438 | 0.368 | 0.500 |
| 188 | 0.650 | 0.599 | 0.707 | 0.541 | 0.438 | 0.604 |
| 250 | 0.628 | 0.567 | 0.687 | 0.565 | 0.510 | 0.619 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 63 | 0.689 | 0.629 | 0.743 | 0.741 | 0.698 | 0.783 |
| 125 | 0.644 | 0.582 | 0.703 | 0.742 | 0.690 | 0.792 |
| 188 | 0.570 | 0.496 | 0.644 | 0.735 | 0.680 | 0.784 |
| 250 | 0.606 | 0.547 | 0.667 | 0.760 | 0.707 | 0.816 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D on IBD: classification performance of Ph-CNN compared to other classifiers on Healthy vs. UCr classification task
| UCr | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 60 | 0.445 | 0.375 | 0.517 | 0.509 | 0.221 | 0.384 |
| 119 | 0.464 | 0.393 | 0.537 | 0.533 | 0.238 | 0.357 |
| 178 | 0.444 | 0.372 | 0.520 | 0.519 | 0.328 | 0.449 |
| 237 | 0.346 | 0.283 | 0.536 | 0.408 | 0.303 | 0.420 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 60 | 0.415 | 0.350 | 0.476 | 0.508 | 0.425 | 0.584 |
| 119 | 0.528 | 0.463 | 0.596 | 0.455 | 0.387 | 0.525 |
| 178 | 0.538 | 0.471 | 0.610 | 0.435 | 0.363 | 0.504 |
| 237 | 0.489 | 0.417 | 0.557 | 0.400 | 0.337 | 0.463 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D on IBD: classification performance of Ph-CNN compared to other classifiers on Healthy vs. CDf classification task
| CDf | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.613 | 0.555 | 0.665 | 0.419 | 0.363 | 0.472 |
| 130 | 0.617 | 0.549 | 0.601 | 0.326 | 0.252 | 0.394 |
| 195 | 0.630 | 0.560 | 0.682 | 0.647 | 0.595 | 0.691 |
| 259 | 0.572 | 0.501 | 0.620 | 0.595 | 0.545 | 0.642 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.610 | 0.549 | 0.666 | 0.677 | 0.618 | 0.728 |
| 130 | 0.620 | 0.551 | 0.685 | 0.706 | 0.648 | 0.758 |
| 195 | 0.601 | 0.534 | 0.667 | 0.739 | 0.685 | 0.788 |
| 259 | 0.648 | 0.589 | 0.703 | 0.720 | 0.667 | 0.768 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D on IBD: classification performance of Ph-CNN compared to other classifiers on Healthy vs. CDr classification task
| CDr | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.241 | 0.172 | 0.311 | 0.138 | 0.073 | 0.198 |
| 129 | 0.232 | 0.167 | 0.295 | 0.089 | 0.028 | 0.151 |
| 193 | 0.202 | 0.131 | 0.273 | 0.169 | 0.101 | 0.236 |
| 257 | 0.218 | 0.158 | 0.278 | 0.178 | 0.107 | 0.251 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.235 | 0. | 0.306 | 0.488 | 0.437 | 0.541 |
| 129 | 0.275 | 0.199 | 0.348 | 0.432 | 0.373 | 0.485 |
| 193 | 0.243 | 0.172 | 0.315 | 0.402 | 0.341 | 0.464 |
| 257 | 0.233 | 0.160 | 0.305 | 0.398 | 0.331 | 0.464 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D on IBD: classification performance of Ph-CNN compared to other classifiers on Healthy vs. iCDf classification task
| iCDf | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 62 | 0.704 | 0.655 | 0.753 | 0.534 | 0.484 | 0.583 |
| 124 | 0.702 | 0.642 | 0.760 | 0.414 | 0.346 | 0.482 |
| 186 | 0.680 | 0.614 | 0.738 | 0.662 | 0.605 | 0.718 |
| 247 | 0.681 | 0.614 | 0.739 | 0.561 | 0.507 | 0.621 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 62 | 0.679 | 0.622 | 0.739 | 0.787 | 0.746 | 0.831 |
| 124 | 0.690 | 0.634 | 0.743 | 0.811 | 0.766 | 0.854 |
| 186 | 0.685 | 0.630 | 0.742 | 0.791 | 0.741 | 0.836 |
| 247 | 0.708 | 0.652 | 0.764 | 0.775 | 0.730 | 0.820 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D on IBD: classification performance of Ph-CNN compared to other classifiers on Healthy vs. iCDr classification task
| iCDr | Ph-CNN | LSVM | ||||
|---|---|---|---|---|---|---|
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 65 | 0.537 | 0.480 | 0.601 | 0.338 | 0.277 | 0.409 |
| 129 | 0.522 | 0.453 | 0.595 | 0.319 | 0.254 | 0.385 |
| 193 | 0.556 | 0.492 | 0.617 | 0.377 | 0.315 | 0.437 |
| 257 | 0.477 | 0.411 | 0.544 | 0.438 | 0.378 | 0.492 |
| MLPNN | RF | |||||
|
| MCC | min CI | max CI | MCC | min CI | max CI |
| 63 | 0.526 | 0.475 | 0.581 | 0.552 | 0.492 | 0.612 |
| 125 | 0.558 | 0.493 | 0.623 | 0.563 | 0.516 | 0.609 |
| 188 | 0.459 | 0.388 | 0.527 | 0.566 | 0.516 | 0.616 |
| 250 | 0.529 | 0.462 | 0.598 | 0.539 | 0.482 | 0.596 |
The performance measure is MCC, with 95% studentized bootstrap confidence intervals (min CI, max CI). Models are computed for p={25%,50%,75% and 100%} of total number of features for each task. Comparing algorithms are linear Support Vector Machines (LSVM), Random Forest (RF) and MultiLayer Perceptron (MLPNN)
Dataset D on IBD: classification performance of Ph-CNN compared to other classifiers on the external validation dataset
| Task | Ph-CNN | LSVM | MLPNN | RF |
|---|---|---|---|---|
| UCf | 0.741 | 0.740 | 0.666 | 0.699 |
| UCr | 0.583 | 0.497 | 0.608 | 0.678 |
| CDf | 0.858 | 0.642 | 0.705 | 0.707 |
| CDr | 0.853 | 0.654 | 0.654 | 0.597 |
| iCDf | 0.842 | 0.418 | 0.401 | 0.920 |
| iCDr | 0.628 | 0.414 | 0.414 | 0.418 |
Fig. 7t-SNE projections of the original features at initial layer (subfigure a) and after 3, 6, 9, 11, 12 convolutional filters (subfigures b-f). Green for healthy subjects, red for iCDf patients