| Literature DB >> 30294269 |
Kai Qiao1, Chi Zhang1, Linyuan Wang1, Jian Chen1, Lei Zeng1, Li Tong1, Bin Yan1.
Abstract
In neuroscience, all kinds of computation models were designed to answer the open question of how sensory stimuli are encoded by neurons and conversely, how sensory stimuli can be decoded from neuronal activities. Especially, functional Magnetic Resonance Imaging (fMRI) studies have made many great achievements with the rapid development of deep network computation. However, comparing with the goal of decoding orientation, position and object category from human fMRI in visual cortex, accurate reconstruction of image stimuli is a still challenging work. Current prevailing methods were composed of two independent steps, (1) decoding intermediate features from human fMRI and (2) reconstruction using the decoded intermediate features. The new concept of 'capsule' and 'capsule' based neural network were proposed recently. The 'capsule' represented a kind of structure containing a group of neurons to perform better feature representation. Especially, the high-level capsule's features in the capsule network (CapsNet) contains various features of image stimuli such as semantic class, orientation, location, scale and so on, and these features can better represent the processed information inherited in the fMRI data collected in visual cortex. In this paper, a novel CapsNet architecture based visual reconstruction (CNAVR) computation model is developed to reconstruct image stimuli from human fMRI. The CNAVR is composed of linear encoding computation from capsule's features to fMRI data and inverse reconstruction computation. In the first part, we trained the CapsNet model to obtain the non-linear mappings from images to high-level capsule's features, and from high-level capsule's features to images again in an end-to-end manner. In the second part, we trained the non-linear mapping from fMRI data of selected voxels to high-level capsule's features. For a new image stimulus, we can use the method to predict the corresponding high-level capsule's features using fMRI data, and reconstruct image stimuli with the trained reconstruction part in the CapsNet. We evaluated the proposed CNAVR method on the open dataset of handwritten digital images, and exceeded about 10% than the accuracy of all existing state-of-the-art methods on the structural similarity index (SSIM). In addition, we explained the selected voxels in specific interpretable image features to prove the effectivity and generalization of the CNAVR method.Entities:
Keywords: brain decoding; capsule network (CapsNet); functional magnetic resonance imaging (fMRI); machine learning; visual reconstruction
Year: 2018 PMID: 30294269 PMCID: PMC6158374 DOI: 10.3389/fninf.2018.00062
Source DB: PubMed Journal: Front Neuroinform ISSN: 1662-5196 Impact factor: 4.081
The corresponding quantitative evaluation for each presented reconstruction in Figure .
| Metrics | a | b | c | d | e | f | g | h | i | j | k | l |
|---|---|---|---|---|---|---|---|---|---|---|---|---|
| MSE | 0.023 | 0.021 | 0.029 | 0.037 | 0.048 | 0.090 | 0.014 | 0.013 | 0.022 | 0.026 | 0.024 | 0.115 |
| PCC | 0.934 | 0.917 | 0.833 | 0.832 | 0.774 | 0.521 | 0.912 | 0.900 | 0.890 | 0.873 | 0.869 | 0.460 |
| SSIM | 0.906 | 0.885 | 0.826 | 0.826 | 0.772 | 0.516 | 0.901 | 0.898 | 0.888 | 0.867 | 0.866 | 0.459 |
The quantitative comparison to other state of the art methods.
| Algorithms | MSE | PCC | SSIM |
|---|---|---|---|
| 0.042 | 0.767 | 0.466 | |
| 0.119 | 0.411 | 0.192 | |
| 0.074 | 0.548 | 0.358 | |
| 0.038 | 0.799 | 0.613 | |
| 0.645 | |||
| Our CNAVR | 0.042 | 0.769 |