| Literature DB >> 30443369 |
Ruben Sanchez-Garcia1, Joan Segura1, David Maluenda1, Jose Maria Carazo1, Carlos Oscar S Sorzano1.
Abstract
Single-particle cryo-electron microscopy (cryo-EM) has recently become a mainstream technique for the structural determination of macromolecules. Typical cryo-EM workflows collect hundreds of thousands of single-particle projections from thousands of micrographs using particle-picking algorithms. However, the number of false positives selected by these algorithms is large, so that a number of different 'cleaning steps' are necessary to decrease the false-positive ratio. Most commonly employed techniques for the pruning of false-positive particles are time-consuming and require user intervention. In order to overcome these limitations, a deep learning-based algorithm named Deep Consensus is presented in this work. Deep Consensus works by computing a smart consensus over the output of different particle-picking algorithms, resulting in a set of particles with a lower false-positive ratio than the initial set obtained by the pickers. Deep Consensus is based on a deep convolutional neural network that is trained on a semi-automatically generated data set. The performance of Deep Consensus has been assessed on two well known experimental data sets, virtually eliminating user intervention for pruning, and enhances the reproducibility and objectivity of the whole process while achieving precision and recall figures above 90%.Entities:
Keywords: cryo-EM; deep learning; image processing; particle pruning; three-dimensional reconstruction
Year: 2018 PMID: 30443369 PMCID: PMC6211526 DOI: 10.1107/S2052252518014392
Source DB: PubMed Journal: IUCrJ ISSN: 2052-2525 Impact factor: 4.769
Figure 1Deep Consensus workflow. Deep Consensus takes the coordinates proposed by different particle pickers as input, from which the intersection (AND set) and the union (OR set) of these coordinates are computed. Next, it picks random coordinates providing that they do not overlap with the OR set (NEG set). The NEG and AND sets are then used to train a convolutional neural network (CNN) that will finally classify the coordinates of the OR set (which is the largest set) as positive particle coordinates or negative particle coordinates.
The deep convolutional neural network architecture employed in Deep Consensus
| Layer No. | Layer type | Kernel size/step size | Shape |
|---|---|---|---|
| 1 | Input | —/— | 128 × 128 × 1 |
| 2 | Conv2D + relu | 15/1 | 128 × 128 × 8 |
| 3 | Conv2D + batch normalization + relu | 15/1 | 128 × 128 × 8 |
| 4 | MaxPooling2D | 7/2 | 64 × 64 × 8 |
| 5 | Conv2D + relu | 7/1 | 64 × 64 × 8 |
| 6 | Conv2D + batch normalization + relu | 7/1 | 64 × 64 × 16 |
| 7 | MaxPooling2D | 5/2 | 32 × 32 × 16 |
| 8 | Conv2D + relu | 3/1 | 32 × 32 × 32 |
| 9 | Conv2D + batch normalization + relu | 3/1 | 32 × 32 × 32 |
| 10 | MaxPooling2D | 3/2 | 16 × 16 × 32 |
| 11 | Conv2D + relu | 3/1 | 16 × 16 × 64 |
| 12 | Conv2D + batch normalization + relu | 3/1 | 16 × 16 × 64 |
| 13 | AveragePooling2D | 4/2 | 8 × 8 × 64 |
| 14 | FullyConnected + relu + dropout ( | 512/— | 512 |
| 15 | SoftMax | 512/— | 2 |
Statistical measurements of Deep Consensus and the performance of the input particle pickers
MCC, Matthews correlation coefficient; ACC, accuracy; ROC-auc, area under the ROC curve; PR-auc, area under the precision-recall curve; NA, not available; DC, Deep Consensus; Xmipp, Xmipp autopicker; Gaussian, EMAN2/SPARX Gaussian picker. MCC, precision and recall were computed at the threshold that maximizes the MCC for DC and at the suggested threshold for Xmipp and Gaussian.
| Algorithm | EMPIAR data set | MCC | ACC | Precision | Recall | ROC-auc | PR-auc |
|---|---|---|---|---|---|---|---|
| DC | 10061 | 0.889 | 0.944 | 0.927 | 0.965 | 0.982 | 0.973 |
| DC | 10028 | 0.942 | 0.971 | 0.958 | 0.984 | 0.993 | 0.989 |
| Xmipp | 10061 | 0.782 | 0.893 | 0.898 | 0.849 | NA | NA |
| Xmipp | 10028 | 0.818 | 0.908 | 0.872 | 0.937 | NA | NA |
| Gaussian | 10061 | 0.697 | 0.845 | 0.778 | 0.904 | NA | NA |
| Gaussian | 10028 | 0.726 | 0.871 | 0.798 | 0.860 | NA | NA |
Figure 2Deep Consensus precision-recall and ROC curves computed from testing sets. Red, the EMPIAR 10028 data set (ribosome); blue, the EMPIAR 10061 data set (β-galactosidase). The area under the curve (auc) is given in parentheses.
Resolution achieved in both data sets when refining different sets of particles
OR, particles selected by any picker; AND, particles selected by both pickers; BPP, particles selected by the picker that obtained the best results; DC-pruned, particles ruled out by Deep Consensus; DC-retained, particles selected as good by Deep Consensus; R, resolution; N, number of particles.
| OR | AND | BPP |
|
| ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| EMPIAR data set |
|
|
|
|
|
|
|
|
|
|
| 10061 | 3.76 | 231251 | 3.32 | 25600 | 2.92 | 117047 | 2.83 | 125586 | 12.74 | 105665 |
| 10028 | 3.83 | 119171 | 3.87 | 67043 | 3.70 | 97561 | 3.65 | 88622 | 33.50 | 30549 |
Figure 3Averages of the major classes obtained from the particles in the AND, OR and DC-retained data sets, all computed using the RELION two-dimensional classification algorithm.
Comparison of different pruning approaches
DC-retained, particles selected as good by Deep Consensus; Z-score-retained, particles that were selected as good using Xmipp particle sorting; R-2D-retained, particles that were selected as good by an expert after using RELION two-dimensional classification; R, resolution; PR, percentage of retained particles compared with the total number of particles (231 251 and 119 171, respectively); T, running time of the pruning algorithm; TT, total running time for the pruning and RELION auto-refine steps.
|
|
| R-2D-retained | ||||||||||
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| EMPIAR data set |
| PR (%) |
| TT (h) |
| PR (%) |
| TT (h) |
| PR (%) |
| TT (h) |
| 10061 | 2.83 | 54.3 | 3.9 | 20.3 | 3.72 | 95.2 | 2.0 | 25.1 | 2.80 | 56.9 | 23.1 | 40.2 |
| 10028 | 3.65 | 74.4 | 2.7 | 12.0 | 3.77 | 94.9 | 1.4 | 14.6 | 3.65 | 85.6 | 14.8 | 27.6 |
Deep Consensus performance on testing sets when trained using synthetic AND sets with different levels of mislabeling noise
R, ribosome data set (EMPIAR-10028); G, β-galactosidase data set (EMPIAR-10061); MCC, Matthews correlation coefficient; ACC, accuracy; ROC-auc, area under the ROC curve. MCC, precision and recall were computed at the threshold that maximizes the MCC.
| MCC | Precision | Recall | ACC | ROC-auc | ||||||
|---|---|---|---|---|---|---|---|---|---|---|
| Corruption level | R | G | R | G | R | G | R | G | R | G |
| 0% | 0.934 | 0.884 | 0.965 | 0.937 | 0.952 | 0.927 | 0.982 | 0.961 | 0.992 | 0.969 |
| 25% | 0.906 | 0.861 | 0.946 | 0.928 | 0.948 | 0.914 | 0.965 | 0.950 | 0.987 | 0.967 |
| 30% | 0.875 | 0.851 | 0.923 | 0.918 | 0.942 | 0.914 | 0.949 | 0.944 | 0.978 | 0.964 |
| 40% | 0.722 | 0.672 | 0.845 | 0.834 | 0.872 | 0.786 | 0.870 | 0.849 | 0.926 | 0.878 |
| 45% | 0.435 | 0.264 | 0.710 | 0.652 | 0.736 | 0.527 | 0.720 | 0.630 | 0.776 | 0.663 |
| 50% | 0.087 | 0.077 | 0.563 | 0.574 | 0.598 | 0.646 | 0.536 | 0.531 | 0.523 | 0.491 |
Deep Consensus precision and recall on testing sets when trained using synthetic AND sets of different sizes with different levels of mislabelling noise
R, ribosome data set (EMPIAR-10028); G, β-galactosidase data set (EMPIAR-10061); #Partic, number of true particles included in the data set; corrupt, corruption level. Each cell displays the precision and recall measured in each condition.
| #Partic | 3000 | 2000 | 1000 | 500 | ||||
|---|---|---|---|---|---|---|---|---|
| Corrupt | R | G | R | G | R | G | R | G |
| 30% | 0.923/0.942 | 0.918/0.914 | 0.920/0.923 | 0.902/0.917 | 0.866/0.926 | 0.876/0.952 | 0.800/0.888 | 0.816/0.900 |
| 40% | 0.839/0.902 | 0.840/0.897 | 0.774/0.819 | 0.835/0.890 | 0.738/0.820 | 0.762/0.595 | 0.693/0.818 | 0.726/0.586 |
| 45% | 0.705/0.762 | 0.695/0.817 | 0.662/0.698 | 0.610/0.701 | 0.602/0.777 | 0.581/0.731 | 0.625/0.709 | 0.578/0.604 |