| Literature DB >> 33175839 |
Ran Li1, Liangyong Yu2, Bo Zhou3, Xiangrui Zeng2, Zhenyu Wang2, Xiaoyan Yang2, Jing Zhang4, Xin Gao5, Rui Jiang1, Min Xu2.
Abstract
Cryo-electron tomography (cryo-ET) provides 3D visualization of subcellular components in the near-native state and at sub-molecular resolutions in single cells, demonstrating an increasingly important role in structural biology in situ. However, systematic recognition and recovery of macromolecular structures in cryo-ET data remain challenging as a result of low signal-to-noise ratio (SNR), small sizes of macromolecules, and high complexity of the cellular environment. Subtomogram structural classification is an essential step for such task. Although acquisition of large amounts of subtomograms is no longer an obstacle due to advances in automation of data collection, obtaining the same number of structural labels is both computation and labor intensive. On the other hand, existing deep learning based supervised classification approaches are highly demanding on labeled data and have limited ability to learn about new structures rapidly from data containing very few labels of such new structures. In this work, we propose a novel approach for subtomogram classification based on few-shot learning. With our approach, classification of unseen structures in the training data can be conducted given few labeled samples in test data through instance embedding. Experiments were performed on both simulated and real datasets. Our experimental results show that we can make inference on new structures given only five labeled samples for each class with a competitive accuracy (> 0.86 on the simulated dataset with SNR = 0.1), or even one sample with an accuracy of 0.7644. The results on real datasets are also promising with accuracy > 0.9 on both conditions and even up to 1 on one of the real datasets. Our approach achieves significant improvement compared with the baseline method and has strong capabilities of generalizing to other cellular components.Entities:
Year: 2020 PMID: 33175839 PMCID: PMC7682871 DOI: 10.1371/journal.pcbi.1008227
Source DB: PubMed Journal: PLoS Comput Biol ISSN: 1553-734X Impact factor: 4.475
Fig 1The flowchart of our method.
Suppose we have a support set with three classes and three labeled samples of each class. Firstly, each support sample is mapped into a task-agnostic embedding space through a 3D encoder and the prototype of each class is calculated. Then a task-specific embedding space is generated through a transformer to focus more on the current classification task, with another set of prototypes calculated. The query sample x is mapped to both embedding spaces respectively and the distances between x and prototypes in both embedding spaces are combined as the classification criteria using a nearest neighbor classifier.
Fig 2An example of the simulated dataset.
(a) Atomic structure of ferritin (PDB ID: 1LB3). (b) Examples of simulated subtomograms containing ferritin macromolecule (PDB ID: 1LB3), represented by several slices of one subtomogram (40 × 40 × 40) in the simulated dataset with SNR = 0.1, 0.05 and 0.03.
Fig 3Architectures of our ProtoNet-CE network.
Details of the 3D encoder and the transformer.
The classification accuracy of the simulated datasets.
5-shot is short for 5-way 5-shot and 1-shot is short for 5-way 1-shot. The suffix (mix) means that the model is trained on a dataset with mixed SNR.
| Methods | SNR = 0.1 | SNR = 0.05 | SNR = 0.03 | |||
|---|---|---|---|---|---|---|
| 5-shot | 1-shot | 5-shot | 1-shot | 5-shot | 1-shot | |
| ProtoNet-CE | 0.8612±0.0165 | 0.7644±0.0216 | 0.7868±0.0194 | 0.7040±0.0214 | 0.6932±0.0212 | 0.5696±0.0205 |
| ProtoNet3D | 0.8432±0.0198 | 0.7480±0.0203 | 0.7567±0.0200 | 0.6901±0.0236 | 0.6631±0.0177 | 0.5287±0.0192 |
| ProtoNet-CE(mix) | 0.8580±0.0185 | 0.7512±0.0198 | ||||
| ProtoNet3D (mix) | 0.7689±0.0253 | 0.7972±0.0201 | 0.6808±0.0236 | 0.6304±0.0212 | ||
| Baseline (fine-tune) | 0.7658±0.0172 | 0.5894±0.0215 | 0.7181±0.0232 | 0.4349±0.0225 | 0.6039±0.0184 | 0.4039±0.0201 |
The classification accuracy of the simulated datasets with different embedding distance used.
| Distance | SNR = 0.1 | SNR = 0.05 | SNR = 0.03 | |||
|---|---|---|---|---|---|---|
| 5-shot | 1-shot | 5-shot | 1-shot | 5-shot | 1-shot | |
| d+d* | 0.7644 | |||||
| d | 0.8432 | 0.7480 | 0.7567 | 0.6901 | 0.6631 | 0.5287 |
| d* | 0.8428 | 0.7648 | 0.6500 | 0.6736 | 0.5524 | |
The classification accuracy of the simulated datasets with different settings of mix training strategy on ProtoNet3D.
Single means the model trained on the dataset with single SNR. Mix100 means the model trained on the dataset with 100 samples for each SNR level. And Mix34 means the model trained on the dataset with 34,33 and 33 samples with SNR = 0.1, 0.05, and 0.03, respectively.
| Methods | SNR = 0.1 | SNR = 0.05 | SNR = 0.03 | |||
|---|---|---|---|---|---|---|
| 5-shot | 1-shot | 5-shot | 1-shot | 5-shot | 1-shot | |
| Single | 0.8432±0.0198 | 0.7480±0.0203 | 0.7567±0.0200 | 0.6631±0.0177 | 0.5287±0.0192 | |
| Mix100 | 0.6808±0.0236 | |||||
| Mix34 | 0.8483±0.0168 | 0.7356±0.0224 | 0.7793±0.0185 | 0.6529±0.0235 | 0.7007±0.0197 | 0.5915±0.0240 |
The classification accuracy of the real datasets of subtomograms.
| Dataset | Methods | 3-way 5-shot | 3-way 1-shot |
|---|---|---|---|
| Guo | ProtoNet3D | 0.9227±0.0076 | 0.8407±0.0153 |
| Guo | ProtoNet-CE | ||
| Guo | Baseline (fine-tune) | 0.8000±0.0135 | 0.5849±0.0152 |
| Noble | ProtoNet3D | ||
| Noble | ProtoNet-CE | ||
| Noble | Baseline (fine-tune) | 0.8702±0.0208 | 0.7965±0.0236 |
Fig 4The results of subtomogram averaging.
(a) Averaged subtomograms after classification. (b) Examples of original subtomograms (before classification) compared to averaged subtomograms (shown in 2D slices).