| Literature DB >> 35462722 |
Abstract
While neural networks can provide high predictive performance, it was a challenge to identify the salient features and important feature interactions used for their predictions. This represented a key hurdle for deploying neural networks in many biomedical applications that require interpretability, including predictive genomics. In this paper, linearizing neural network architecture (LINA) was developed here to provide both the first-order and the second-order interpretations on both the instance-wise and the model-wise levels. LINA combines the representational capacity of a deep inner attention neural network with a linearized intermediate representation for model interpretation. In comparison with DeepLIFT, LIME, Grad*Input and L2X, the first-order interpretation of LINA had better Spearman correlation with the ground-truth importance rankings of features in synthetic datasets. In comparison with NID and GEH, the second-order interpretation results from LINA achieved better precision for identification of the ground-truth feature interactions in synthetic datasets. These algorithms were further benchmarked using predictive genomics as a real-world application. LINA identified larger numbers of important single nucleotide polymorphisms (SNPs) and salient SNP interactions than the other algorithms at given false discovery rates. The results showed accurate and versatile model interpretation using LINA.Entities:
Keywords: Interpretable machine learning; bioinformatics; deep neural networks; predictive genomics
Year: 2022 PMID: 35462722 PMCID: PMC9032252 DOI: 10.1109/access.2022.3163257
Source DB: PubMed Journal: IEEE Access ISSN: 2169-3536 Impact factor: 3.476
FIGURE 1.First-order model-wise interpretation. The three bars of a feature represented the FP, IP and DP scores of this feature in the LINA model.
FIGURE 2.Second-order model-wise interpretation. The second-order model-wise importance scores (SP) are undirected between two features and are shown in a symmetric matrix as a heatmap. The importance scores for the feature self-interactions are set to zero in the diagonal of the matrix.
Benchmarking of the first-order interpretation performance using five synthetic datasets (F1~F5)*
| Methods╲Datasets | F1 | F2 | F3 | F4 | F5 | Average |
|---|---|---|---|---|---|---|
|
|
| 0.88±0.03 | 0.25±0.07 | 0.65±0.05 | 0.92±0.03 | 0.74±0.04 |
|
|
| 0.92±0.03 | 0.69±0.01 | 0.84±0.03 | 0.96±0.03 | 0.88±0.02 |
|
|
| 0.97±0.02 |
|
|
|
|
|
| 0.99±0.01 |
| 0.95±0.03 | 0.83±0.12 |
| 0.95±0.03 |
|
|
| 0.90±0.01 |
| 0.76±0.11 |
| 0.93±0.03 |
|
|
|
| 0.85±0.08 | 0.78±0.12 |
| 0.93±0.04 |
|
| 0.59±0.06 | 0.41±0.07 | 0.15±0.11 | 0.30±0.08 | 0.5±0.03 | 0.39±0.07 |
|
| −0.72±0.0 | −0.52±0.08 | −0.14±0.07 | −0.57±0.05 | −0.3±0.06 | −0.45±0.05 |
The best Spearman correlation coefficient for each synthetic dataset is highlighted in bold
Precision of the second-order interpretation by LINA SP, NID and GEH in ten synthetic datasets (F6 F10).
| Total Features | Datasets | NID | GEH | LINA SP |
|---|---|---|---|---|
|
|
| 44.5% ±0.2% | 50.0%±0.2% |
|
|
|
| 41.0%±0.2% | 92.0%±0.1% | |
|
| 80.6%±0.2% | 48.8%±0.4% |
| |
|
| 62.2%±0.4% | 41.4%±0.3% |
| |
|
| 56.7%±0.3% | 75.0%±0.5% |
| |
|
| 68.4%±0.2% | 51.2%±0.3% |
| |
|
|
| 51.8%±0.2% | 18.1%±1.0% |
|
|
| 44.0%±0.2% | 28.8%±0.4% |
| |
|
| 76.3%±0.1% | 47.9%±0.2% |
| |
|
| 40.0%±0.3% | 41.8%±0.2% |
| |
|
|
| 10.4%±1.0% | 50.0%±0.1% | |
|
| 55.7%±0.2% | 29.4%±0.6% | 64.9%±0.2% |
The best precision for each dataset is highlighted in bold
Performance benchmarking of the first-order interpretation for predictive genomics.
| Methods | LINA FP | Saliency | grad*Input | DeepLift | LIME | L2X |
|---|---|---|---|---|---|---|
|
|
| 35 | 75 | 75 | 9 | 0 |
|
|
| 35 | 88 | 85 | 9 | 0 |
|
|
| 57 | 122 | 119 | 9 | 0 |
|
|
| 7.5% | 3.0% | 2.0% | 16.3% | N/A |
|
|
| 16.2% | 9.3% | 9.3% | 20.5% | N/A |
Performance benchmarking of the second-order interpretation for predictive genomics.
| Methods | LINA SP | NID | GEH |
|---|---|---|---|
|
|
| 415 | 0 |
|
|
| 504 | 0 |
|
|
| 810 | 0 |
|
|
| 10.5% | N/A |
|
|
| 31.8% | N/A |
| Order | Target | Acronym | Definition |
|---|---|---|---|
| First-order | Instance-wise | FQ | FQi = DQi + IQi |
| DQ | DQi = | ||
| IQ |
| ||
| Model-wise | FP |
| |
| DP |
| ||
| IP |
| ||
| Second-order | Instance-wise | SQ |
|
| Model-wise | SP |
|