Literature DB >> 35759502

Comparing the reliability of different ICA algorithms for fMRI analysis.

Pengxu Wei1,2, Ruixue Bao3, Yubo Fan2.   

Abstract

Independent component analysis (ICA) has been shown to be a powerful blind source separation technique for analyzing functional magnetic resonance imaging (fMRI) data sets. ICA can extract independent spatial maps and their corresponding time courses from fMRI data without a priori specification of time courses. Some popular ICA algorithms such as Infomax or FastICA generate different results after repeated analysis from the same data volume, which is generally acknowledged as a drawback for ICA approaches. The reliability of some ICA algorithms has been explored by methods such as ICASSO and RAICAR (ranking and averaging independent component analysis by reproducibility). However, the exact algorithmic reliability of different ICA algorithms has not been examined and compared with each other. Here, the quality index generated with ICASSO and spatial correlation coefficients were used to examine the reliability of different ICA algorithms. The results demonstrated that Infomax running 10 times with ICASSO could generate consistent independent components from fMRI data sets.

Entities:  

Mesh:

Year:  2022        PMID: 35759502      PMCID: PMC9236259          DOI: 10.1371/journal.pone.0270556

Source DB:  PubMed          Journal:  PLoS One        ISSN: 1932-6203            Impact factor:   3.752


Introduction

Independent component analysis (ICA) has been shown to be a powerful blind source separation technique for analyzing functional magnetic resonance imaging (fMRI). ICA can decompose a set of signal mixtures into a corresponding set of statistically independent component signals (source signals) [1-4]. The basic concept of ICA can be expressed as X = M ∙ C, where X is the observed data (i.e., the data matrix of fMRI signal), C is the component map (a matrix of voxel values), and M is the mixing matrix determining the time-varying contribution of each component map to the observed fMRI data. To estimate M and C simultaneously, ICA determines the unmixing matrix W (a permuted version of the inverse of the mixing matrix M) through iterative calculation. The component maps and corresponding time courses can be acquired using the following equation: C = W ∙ X [5, 6]. A number of ICA algorithms have been used for fMRI data analysis, including Infomax (Information Maximization), FastICA, ERICA (Equivalent Robust Independent Component Analysis), and SIMBEC (SIMultaneous Blind signal Extraction using Cumulants). ICA can extract independent spatial maps and their corresponding time courses from fMRI data without a priori specification of time courses. However, most ICA algorithms are based on minimization or maximization of an objective function and will find different local minima depending on where the “initial point” is. The randomness of the initialization that inherently exists in such ICA algorithms introduces randomness into the ICA decomposition and leads to different results when running ICA repeatedly. As a result, some popular ICA algorithms such as Infomax or FastICA will generate different results after repeated analysis from the same data volume, which is generally acknowledged as a drawback for ICA approaches [7, 8]. The reliability of some ICA algorithms has been explored by methods such as RAICAR (Ranking and Averaging Independent Component Analysis by Reproducibility) [7] and ICASSO (http://research.ics.aalto.fi/ica/icasso/) [8]. Such methods repeat the analysis multiple times with different random initializations and subsequently quantify the consistency of the outcomes [7-13]. ICASSO provides several indices including the quality index (Iq) to assess the reliability of independent components (ICs) after multiple runs. The Iq is introduced to be the main index reflecting the compactness and isolation of a cluster that contains ICs with high similarities from multiple runs. The Iq is computed as the difference between the average intra-cluster similarities and average extra-cluster similarities. RAICAR uses the spatial correlation coefficients (SCCs) to measure the component reproducibility. The reliability of the FastICA algorithm was assessed by using RAICAR or its extensions/variaions [7-10], and Correa et al. compared the results from Infomax, FastICA, JADE (Joint Approximate Diagonalization of Eigenmatrices), and EVD (Eigenvalue Decomposition)and found inconsistency between the results of EVD [14]. Some issues regarding the reliability of different ICA algorithms still remain. (1) The exact algorithmic reliability of different ICA algorithms is still unknown. Yang et al. assessed the algorithmic reliability of FastICA by using SCC [7]. Correa et al. applied ICASSO to evaluate the consistency of results from multiple runs of FastICA by using Infomax as a benchmark; however, they did not measure the reliability of Infomax [14]. No published reports examined the difference of different ICA algorithms with quantified index such as SCC (used by RAICAR) or Iq (provided by ICASSO). (2) ICA algorithms including AMUSE (Algorithm for Multiple Unkown Signal Extraction), JADE, ERICA (Extended Robust Independent Components Analysis), RADICAL (Robust, Accurate, Direct ICA aLgorithm), and SIMBEC produce identical results after each run (deterministic algorithms), whereas Infomax, FastICA, EVD, and COMBI (Combination of WASOBI and EFICA; WASOBI: Weights-Adjusted Second Order Blind Identification, EFICA: Efficient variant of algorithm FastICA) generate different results after repeated analysis (non-deterministic algorithms). How many times a non-deterministic algorithm should be run to acquire a reliable result remains unknown. Yang et al. [7] found that more than 20 times was not necessary, but only FastICA was tested. ICASSO presents clusters after integrating the results of multiple runs. The estimates of an IC will be incorporated into a cluster if these estimates are consistent, and the number of estimates in a cluster is equivalent to the number of runs if the corresponding ICs are estimated consistently. Therefore, the number of estimates in each cluster can also be used to assess the reliability of an ICA algorithm. This study aimed to solve three problems. (1) The Iq and the number of estimates in each cluster provided by ICASSO were used together to quantify the reliability of four non-deterministic algorithms (i.e., Infomax, FastICA, EVD, and COMBI). (2) Each of the four non-deterministic algorithms was repeated at different times with ICASSO (from 10 times to 100 times for each algorithm), and the SCC was used to determine how many times were needed to acquire ICA results with good reliability. (3) The SCC was used to quantify the consistency among the results of nine ICA algorithms (the four non-deterministic algorithms and the five deterministic algorithms, including AMUSE, JADE, ERICA, RADICAL, and SIMBEC) and determine the algorithm with the best reliability. We explored real fMRI data involving sensory stimulation and motor task execution in this study.

Materials and methods

Subjects, stimulation/task paradigms, and image acquisition

The sensory stimulation data were acquired by using a 1.5T MRI scanner from a female patient (aged 18 years) with traumatic brain injury. The fMRI data were acquired with a gradient echo type echo planar imaging (GRE–EPI) sequence (Repetition Time = 3000 ms, Echo Time = 40 ms, flip angle = 90°, field of view = 240 mm × 240 mm, matrix size = 64 × 64, slice thickness = 5 mm, no gap, 23 contiguous axial slices). In one fMRI session, a block design protocol was applied with successive blocks alternating between rest and transcutaneous electrical nerve stimulation, starting with rest. There was a 24 s dummy period before MR data collection. Each block duration of rest or electrical stimulation was 30 s. The rest–stimulation cycle was repeated six times. Thus, the total duration of the fMRI session was 6 min 24 s. The electrical stimulation was performed at a rate of 2 Hz with surface electrodes. The cathode was placed on the space between the spinous processes of L2 and L3, whereas the anode was placed on the space between the spinous processes of L4 and L5. The stimulation was performed with a WQ-10D stimulator (Beijing Electronics Instrument Co., Ltd.). The intensity was set to a score of 2 rated on a 0–10 verbal numerical rating scale. T1-weighted images were also acquired to provide an anatomical reference. The motor task data were acquired by using the same scanner from a healthy male subject (aged 30 years). The fMRI data were acquired with a GRE–EPI sequence (Repetition Time = 3000 ms, Echo Time = 40 ms, flip angle = 90°, field of view = 240 mm × 240 mm, matrix size = 64 × 64, slice thickness = 5 mm, no gap, 28 contiguous axial slices). T1-weighted images were also acquired. Two types of motor tasks were applied (i.e., imagined movements and motor execution task) in one fMRI session. The motor execution task consisted of 90° isometric dorsiflexion of the right foot. During the imagined movement task, the subject was required to imagine performing the same type of foot movement instead of actually executing it. The total duration of the fMRI session was 6 min. The experiment consisted of eight 21 s periods of baseline alternating with eight 21 s periods of motor tasks. The imagined movements were followed by the motor execution task in sequence. The beginning and ending of each task block were signaled with verbal commands “Ready, imaging, right, start” and “stop” or “Ready, move, right, start” and “stop.” The duration of each command was 3 s. An audio cue repeated every 3 s was broadcasted during each task block to prompt the subject to persistently focus on the task. The two fMRI data sets contained three types of status, i.e., sensory stimulation, imagined movements and motor execution task. The fMRI data sets can be found in the S1 Data. The experiment was conducted with the approval from the institutional review board of the National Research Center for Rehabilitation Technical Aids, and written informed consents were obtained from the subjects. The procedure of the experiment was in accordance with the principles of the Declaration of Helsinki.

Performing ICA

The Statistical Parametric Mapping (SPM) software (https://www.fil.ion.ucl.ac.uk/spm) was used to perform motion correction, coregistration, and spatial smoothing with a Gaussian kernel of 10 mm full-width at half-maximum. ICA was performed using the Group ICA of fMRI Toolbox (GIFT) software (http://mialab.mrn.org/software/gift/index.html). The number of ICs was estimated by using the minimum description length (MDL) criteria [15]. Except for the algorithm and the number of ICs, default settings/parameters defined by the GIFT software were used during analysis. In group analysis containing a number of subjects, subjects in a group can be concatenated as a single ensemble by the GIFT software. In this study, there was a single ensemble for each subject since each data set contained only one subject. During data reduction steps, for one subject and one session, the data reduction actually would be disabled since the number of principal components extracted from the data is the same as the number of independent components.

Assessing the reliability of four non-deterministic ICA algorithms with ICASSO

The ICASSO toolbox embedded in the GIFT software was employed to run 10 times with different repetition number k (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100) for each of the four non-deterministic algorithms. The RandInit mode (algorithm starts with Randomizing different Initial values) was used. This mode in ICASSO uses the original data whereas the data will be resampled if the bootstrapping method is used. Additionally, the RandInit mode generates correlation coefficients with straightforward calculations whereas some extra normalization is necessary for bootstrapping [8]. When we run ICASSO 10 times by using the RandInit mode, the algorithm (e.g., Infomax) would run 10 times; each time the algorithm started with randomizing different initial conditions. For a given algorithm, each of the 10 results of ICASSO contained N clusters; the number of N was determined by the MDL criteria. For an ICASSO result with a repetition number k, clusters presented by the ICASSO toolbox included estimated ICs from each run if these ICs presented high mutual similarities. The centrotype of a cluster holds the maximum sum of similarities to other points in the cluster and was used as the representation of this cluster by ICASSO. In ICASSO, the similarity between one pair of ICs (i and j) is quantified by the absolute value of their mutual correlation coefficients σ. The clustering process is performed by using the distance between the two ICs. The distance is determined by transforming the similarity matrix into a dissimilarity (distance) matrix: d = 1-σ [8]. The reliability of the four non-deterministic algorithms was assessed with two approaches. The Iq was used to be the main index of reliability. First, for each algorithm, we compared the Iq values among results with different repetition number k to determine whether there is any difference when the repetition number k increases and to find the best k value with which a given algorithm could provide the most reliable result. Second, the number of ICs in each cluster will be equal to the repetition number k if an ICA algorithm is reliable, whereas an unreliable algorithm will generate clusters containing less number of ICs than the repetition number k. Thus, we compared how many clusters contained different numbers of ICs from the repetition number k for each of the four non-deterministic algorithms. Afterward, we compared the Iq values between the most reliable results of the four non-deterministic algorithms to determine which ICA algorithm exhibited the highest reliability (i.e., the highest Iq value).

Comparing the reliability among four non-deterministic ICA algorithms with SCC

The SCCs between each pair of ICs from each pair of ICASSO results using the same algorithm were calculated by using MATLAB function corrcoef. The SCC was the Pearson correlation coefficient value between spatial maps of a pair of ICs. Afterward, the SCCs were used in two steps. Step 1 was a process to select the maximum value. For a non-deterministic algorithm, suppose that the total number of ICs in each ICASSO result (when k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100, the corresponding result was A, B, C, …, J, respectively) is i. Then, the j-th IC map (j = 1, 2, …, i) in ICASSO result A would have i SCC values with each IC map in ICASSO result B for this algorithm. The highest one among these i SCC values indicates the best spatial consistency/similarity. Thus, an IC with the highest SCC value in ICASSO result B can be considered to match the j-th IC in ICASSO result A. In this way, each pair of matched ICs in each pair of ICASSO results was determined. As a result, we acquired a list of SCCs for each pair of groups. In Step 2, for each non-deterministic algorithm, the SCC values between the most reliable result (determined in the former section) and the other nine results were compared by using the Kruskal–Wallis test to measure the impact of repetition time k on the consistency of spatial maps.

Comparing the reliability among nine ICA algorithms with SCC

Finally, we compared the results from the nine algorithms with the SCC value as an index to investigate the spatial difference of results acquired with different algorithms. For the non-deterministic ICA algorithms Infomax, FastICA, EVD, and COMBI, the most reliable results were used to represent the most approximation of the source.

Results

Using the MDL criteria, the estimated numbers of ICs were 12 for the sensory stimulation data and 8 for the motor task data.

The most reliable result for the four non-deterministic algorithms when using the Iq value as an index

Sensory stimulation data

The median Iq values of different results for each of the four non-deterministic algorithms are shown in . The fluctuation of Infomax results appeared to be much less than those of the other algorithms. For each of the four algorithms, the ICASSO repetition number k did not influence the Iq values (). For all of the four algorithms, the repetition number k with the highest median is displayed in . The ICASSO results with such number k were the most reliable ones for each non-deterministic algorithm.

Median Iq values of different results for non-deterministic ICA (sensory data).

The unit of the x-axis is the ICASSO repetition number k, and the unit of the y-axis is the Iq value. The results of COMBI presented clusters containing only one estimate when k = 10, 20, 40, and 70; under such conditions, the corresponding Iq value could not be calculated, which suggested very poor reliability. For each non-deterministic algorithm, the Iq values of different repetition number k (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100) did not present statistically significant difference (Kruskal–Wallis test). The result of the Kruskal–Wallis test of COMBI was based on six groups of Iq values because four results (when k = 10, 20, 40, and 70) of COMBI had clusters without the Iq value. The result of COMBI was based on six groups of Iq values because four results (when k = 10, 20, 40, and 70) had clusters without the Iq value.

Motor task data

The median Iq values of different results for each of the four non-deterministic algorithms are shown in . The fluctuation of results of Infomax also appeared to be much less than those of the other three algorithms. For each of the four algorithms, the ICASSO repetition number k did not influence the Iq values (). For all of the four algorithms, the repetition number k with the highest median is displayed in .

Median Iq values of different results for non-deterministic ICA (motor data).

The unit of the x-axis is the ICASSO repetition number k, and the unit of the y-axis is the Iq value. For each non-deterministic algorithm, the Iq values of different repetition number k (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100) did not present statistically significant difference (Kruskal–Wallis test). Taken together, Infomax presented better reliability than the other three non-deterministic algorithms.

Most reliable non-deterministic algorithm when using Iq values as the index

For each non-deterministic algorithm, the ICASSO result with the highest median Iq values was used as the most reliable one (i.e., those shown in for the sensory data and for the motor data). Here, these results were used to represent each algorithm and were then compared. The results of the Kruskal–Wallis test are shown in . For both sensory and motor data, Infomax generated higher median Iq values than other algorithms although the difference was not statistically significant (p = 0.1709, Table 5). For the motor data, the median Iq value of the most reliable results from EVD (k = 80) was significantly lower than those of the other three non-deterministic algorithms. The result of COMBI was significantly lower than those of Infomax and FastICA when comparing the results from Infomax, FastICA, and COMBI (Chi-sq 15.7650, p = 0.0003773, the medians shown in Table 4). The results of Infomax and FastICA were not statistically different (Chi-sq 0.8934, p = 0.3446).
Table 5

Differences among the most reliable results of each non-deterministic algorithm.

Sensory dataMotor data
Chi-sq5.012822.0256
P0.17090.00006
Table 4

ICASSO repetition number k with the highest median Iq of each algorithm (motor data).

InfomaxFastICAEVDCOMBI
k10208020
median0.99540.99020.92420.9756
Thus, Infomax presented better reliability than the other three non-deterministic algorithms.

Assessing the reliability using the number of clusters containing different numbers of ICs from ICASSO repetition number k as an index

For Infomax, all clusters of each ICASSO result contained the same number as the repetition number k; for FastICA, each ICASSO result contained at least two clusters presenting different numbers of ICs from ICASSO repetition number k when k > 30, and the total number of such clusters was 21; for EVD, the ICASSO results contained two clusters presenting different numbers of ICs from the repetition number k when k = 80; for COMBI, each ICASSO result contained at least two clusters presenting different numbers of ICs from the repetition number k, and the total number of such clusters was 48. For Infomax, all clusters of each ICASSO result contained the same number as the repetition number k; for FastICA, when k = 50, 70, and 80, each ICASSO result contained two clusters presenting different numbers of ICs from the repetition number k, and the total number of such clusters was 6; for EVD, when k = 90, the ICASSO results contained two clusters presenting different IC numbers from the repetition number k; for COMBI, when k > 50, each ICASSO result contained at least two clusters presenting different numbers of ICs from the repetition number k, and the total number of such clusters was 13. Taken together, Infomax exhibited the best reliability when using the number of clusters containing IC numbers different from the ICASSO repetition number k as an index.

Difference in SCC values between the most reliable results and the other results for each non-deterministic algorithm

For each algorithm, no statistical difference in SCC values was found between the most reliable ICASSO results and each of the other nine results for either sensory data () or motor data (). For each non-deterministic algorithm, ICASSO was run k times (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, 100) to acquire 10 results. The most reliable ICASSO results (shown in Table 2 for the sensory data) and the other results were compared by using the Kruskal–Wallis test.
Table 2

ICASSO repetition number k with the highest median Iq of each algorithm (sensory data).

InfomaxFastICAEVDCOMBI
k60606060
median0.98430.95220.96670.9190

The result of COMBI was based on six groups of Iq values because four results (when k = 10, 20, 40, and 70) had clusters without the Iq value.

For each non-deterministic algorithm, ICASSO was run k times (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, 100) to acquire 10 results. The most reliable ICASSO results (shown in Table 4 for the motor data) and the other results were compared by using the Kruskal–Wallis test. The range of SCC values for each non-deterministic algorithm is presented in Tables and . Note for the sensory data, FastICA and COMBI had very low SCC values, indicating poor spatial reliability. For the sensory data, if the SCC values ≤ 0.88 (found in the results of COMBI), there would be an IC (in the corresponding ICASSO result) that matched two IC maps of the most reliable result (i.e., presenting similar SCC values), and there was another IC (in the corresponding ICASSO result) that did not match any IC in the most reliable result. In other words, when compared with the most reliable result, the other nine results with SCC ≤ 0.88 indicated unreliable performance. Thus, we set 0.88 as a threshold to numbers of such SCC values in Tables and (using 0.9 referring to [7]). Taken together, Infomax exhibited the best reliability.

Comparing SCC values among the nine ICA algorithms

In previous sections, Infomax always presented better reliability than other non-deterministic algorithms. Infomax is a widely used ICA algorithm. It has been proven to be quite reliable for fMRI data analysis [6, 14, 16] and is a well-performing algorithm for other data types [17, 18]. Thus, we used ICASSO results of Infomax as a reference to compare the performance of other ICA algorithms against Infomax. For either sensory data or motor data, SCC values were computed between Infomax (the most reliable ICASSO result) and the results of the other eight ICA algorithms. For FastICA, EVD, and COMBI, the most reliable ICASSO results were used. For both sensory data and motor data, only the results of FastICA exhibited good spatial consistency with the results of Infomax when considering median and minimum SCC values (Tables and ). Other algorithms had very low minimum SCC values in the results of either sensory or motor data, or both of them. The results of FastICA (the most reliable results) presented higher SCC values than other algorithms. If there was an SCC value less than 0.669816, there would be at least one “unmatched” IC in the result of such algorithm, which means that this IC could not exclusively match any IC map in the results of Infomax and thus suggests a poor spatial consistency between the results of the algorithm and Infomax.

Discussion

Whichever index was used (Iq values, cluster numbers, and SCC values), Infomax always presented better reliability than other non-deterministic algorithms (FastICA, EVD, and COMBI). Infomax also exhibited better consistency than the five deterministic algorithms (AMUSE, JADE, ERICA, RADICAL, and SIMBEC) when using SCC as the index. Compared with Infomax, FastICA presented weaker consistency with some unreliable separated ICs. Other algorithms exhibited poor consistency under some conditions. When running Infomax with ICASSO, repetition times > 10 did not lead to significant differences between results, indicating that 10 times can be used to acquire reliable ICs when using this algorithm. In additional to the high reliability of Infomax, this finding can also be attributed to the algorithm of ICASSO that generates centrotype of each cluster as the representative. The centrotype is calculated as the maximum sum of similarities to other points in a cluster [8], and running Infomax with ICASSO can therefore generate consistent results after a few repetition times. When using ICA to explore cerebral networks, especially for conditions with low signal-to-noise ratio such as resting state [19], consistent ICA results generated by using a reliable algorithm such as Infomax can lead to higher study efficiency. Otherwise, variations of ICA results from different runs enhance the difficulty in explaining acquired ICs. Our results demonstrate that Infomax running 10 times with ICASSO can generate consistent ICs from fMRI data sets. This finding provides an easily accessible approach to generate reliable ICs. These ICs can be used to develop advanced analyses such as personalized brain networks [20]. The results also demonstrate that the algorithms other than Infomax produce more or less unreliable separated ICs. Therefore, caution needs to be taken in explaining ICA results when the consistency of ICs cannot be assured. The SCC was the correlation coefficient value between two IC spatial maps, that is, two spatial matrices. When comparing two groups of ICs A and B, the correlation coefficient value between each IC in group A and each IC in group B was calculated. The best-matched ICs must be a pair presenting the highest correlation value. Such a pair was unique since the corresponding correlation value was the highest one, and we did not find an IC in one group presenting two equally highest correlation values with two different ICs in the other group. We used the correlation coefficient value to compare different conditions, and the correlation values can be transformed into normal variables by using the Fisher-Z transform. Suppose that the population of correlation coefficient values is ρ, and the sample of correlation coefficients is r (the sample is a part of the population). When the Fisher-Z transformation is applied to the sample r, the sampling distribution of the transformed variable is approximately normal. Without the Fisher transformation, the variance of r (between two variables X and Y) grows smaller as |ρ| gets closer to 1 and thus is not normally distributed. In this study, however, for each comparison between two groups of SCCs, the normal distribution of the transformed SCC values in each group cannot be assured. The reason is that the SCC values (r values) come from different populations. For example, the SCC value between j-th IC in group A and k-th IC in group B can be Fisher-Z transformed, which leads to a normal distribution of the transformed SCC value. Such a normal distribution means that the transformed r values meet normal distribution for the two populations where two samples (j-th IC in group A and k-th IC in group B) are drawn. But now we have other SCC values between other pairs of ICs (e.g., m-th IC in group A and n-th IC in group B). These are samples from other populations. Therefore, the transformed r values may not be normally distributed when put into a group. As the solution, we used the Kruskal–Wallis test in Tables 6 and 7 since the Kruskal–Wallis test is a nonparametric test and does not assume the normality of the measurement variable.
Table 6

Differences in SCC values between the most reliable ICASSO results and the other nine results (sensory data).

InfomaxFastICAEVDCOMBI
Chi-sq2.37340.72406.18185.0999
P0.96740.99950.62690.7468

For each non-deterministic algorithm, ICASSO was run k times (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, 100) to acquire 10 results. The most reliable ICASSO results (shown in Table 2 for the sensory data) and the other results were compared by using the Kruskal–Wallis test.

Table 7

Differences in SCC values between the most reliable ICASSO results and the other nine results (motor data).

InfomaxFastICAEVDCOMBI
Chi-sq3.25464.03727.02662.9837
P0.91740.85370.53380.9354

For each non-deterministic algorithm, ICASSO was run k times (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, 100) to acquire 10 results. The most reliable ICASSO results (shown in Table 4 for the motor data) and the other results were compared by using the Kruskal–Wallis test.

This study used fMRI data from two subjects. These fMRI data sets contained three types of status, i.e., sensory stimulation, imagined movements, and motor execution task. Single-subject ICA analysis is the foundation of group-level analysis. However, results from two single-subject analyses may not well represent the performance of an ICA algorithm for all types of data. Whether the reliability of Infomax is superior to other algorithms needs to be further proved with other types of data. (RAR) Click here for additional data file. 25 Jan 2022
PONE-D-21-19322
Comparing the reliability of different ICA algorithms for fMRI analysis
PLOS ONE Dear Dr. Wei,
 
Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Sorry for the long review process so far, and we had only one reviewer's feedback, the experts comments are comprehensive and in-depth. I feel it is fair to move forward with you starting the revision process and I personally went over the MS couple time and concur with the reviewer. Please submit your revised manuscript by Mar 11 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Kewei Chen, Ph.D Academic Editor PLOS ONE Journal Requirements: When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section. 3. Thank you for stating the following in the Acknowledgments Section of your manuscript: "This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700), the National Natural Science Foundation of China (Grant No. 81972160), and the Beijing Natural Science Foundation (Grant No. 17L20019)." We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: "This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700 by WP and ZL), the National Natural Science Foundation of China (Grant No. 81972160 by WP), and the Beijing Natural Science Foundation (Grant No. 17L20019 by WP). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript." Please include your amended statements within your cover letter; we will change the online submission form on your behalf. 4. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability. Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized. Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access. We will update your Data Availability statement to reflect the information you provide in your cover letter. 5. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well. 6. Please include a copy of Table S8, S9, S2 and S4 which you refer to in your text on pages 6, 11 and 12. 7. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 2. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: No ********** 3. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: No ********** 4. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: Yes ********** 5. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: This is an interesting article. There is real value. However, I have several concerns regarding details missing, and technical questions that are unanswered in the article in its present form. The article is generally well written, but all acronyms are not fully spelled out on first use, and software packages and links are not always provided in the methods. The authors would have made both review and their revision tasks easier if they had added line numbers in the draft. I will detail the various data and methodology concerns in detail below. 1. ACRONYMS must be spelled out on first use and links to software and citations provided. The omissions are numerous. A few (not complete list) MDL, AMUSE, JADE, ERICA and surprisingly, even ICASSO which is a pivotal method here. 2. Only two human subject data sets are used in this paper. This might be acceptable if we were convinced that the data was sufficiently rich, but this information is missing. The reader needs to know fully data set sizes: numbers of imaging sessions, full number of tasks, full number of stimuli in tasks etc. The richness of data is a factor in ICA outcomes so these aspects MUST be reported. Similarly, the method of data setup for ICA needs better description. Was data analyzed in each subject ultimately as a single concatenated ensemble (and if so, of what matrix dimensions), or as a number of subset ensembles (and if so, of what dimensions, and how many such ensembles). Because only two individuals data were used, there remains the unfortunate possibility that there are structural individual-based idiosyncracies in the data here that favor Infomax. I do not believe this likely, but it cannot be fully discounted given the data. Other research applying ICA or testing dimensionality separation methods including ICA have used larger numbers of individuals, or/and artificial data generation under constraints to explore comparisons, or reliability. Examples in motor field are Tresch, D'Avella, Cheung, papers in J Neurophysiology for synergy separation analyses, and Yang, Logan and Giszter in PNAS exploring SCC-like measures on motor synergy outcomes across individual animals. An aspect not explored by the authors of this paper (and given lack of detail provided, this may or may not be possible in their data), that might be used to enrich the analysis they perform is bootstrapping or jack-knifing the data sets used, in addition to subsequent multiple iterations of ICA methods on each subjects subset data. At the very least, there needs to be discussion of the data limits in this paper resulting from only 2 subjects and only 1 subject per type of experiment, and the caveats resulting, i.e. possibility of individual idiosyncracy of data favoring specific ICA methods. 3. The SCC method as described in section 2.4 is unclear. It is possible to do this in two ways - correlate the individual IC spatial components picking best correlations (where pairings of and IC with other sets may not be unique- same IC best in two or more correlations), or correlate the spatial matrix (e.g., in MATLAB using matperm, matcorr) and then IC correlations will be unique, based on the matrix permutation use in the unique IC matchings. The authors used the former, non-unique method I believe, but this might have biased results. If they did, with the possibility of non-unique matching of ICs, then I think the reader needs to know if there were non-unique best IC spatial correlations in each method, and if so how many. This might be an additional metric on ICA algorithm quality. 4. Finally, in relation to the SCC statistics used: Correlation, especially here, is non-gaussian if used directly. This may present difficulties in interpreting the statistics in tables 6 and 7 that could be avoided. Using the Fisher-Z transform (see Yang, Logan, Giszter, PNAS 2019 for example with Infomax ICA data stats) the correlations are transformed to normal variables, improving the interpretation of parametric methods. 5. Equations would be helpful throughout the methods if very clearly written. The goal is reproducibility of methods. 6. The data sharing statement is insufficient, for this study an anonymized repository should be chosen - there are many. This paper is a solid contribution, but marred by the omissions in detail, and possibly improved by attention to the technical points noted. ********** 6. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step. 3 Feb 2022 When submitting your revision, we need you to address these additional requirements. 1. Please ensure that your manuscript meets PLOS ONE's style requirements, including those for file naming. The PLOS ONE style templates can be found at https://journals.plos.org/plosone/s/file?id=wjVg/PLOSOne_formatting_sample_main_body.pdf and https://journals.plos.org/plosone/s/file?id=ba62/PLOSOne_formatting_sample_title_authors_affiliations.pdf 2. We note that the grant information you provided in the ‘Funding Information’ and ‘Financial Disclosure’ sections do not match. When you resubmit, please ensure that you provide the correct grant numbers for the awards you received for your study in the ‘Funding Information’ section. 3. Thank you for stating the following in the Acknowledgments Section of your manuscript: "This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700), the National Natural Science Foundation of China (Grant No. 81972160), and the Beijing Natural Science Foundation (Grant No. 17L20019)." We note that you have provided funding information that is not currently declared in your Funding Statement. However, funding information should not appear in the Acknowledgments section or other areas of your manuscript. We will only publish funding information present in the Funding Statement section of the online submission form. Please remove any funding-related text from the manuscript and let us know how you would like to update your Funding Statement. Currently, your Funding Statement reads as follows: "This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700 by WP and ZL), the National Natural Science Foundation of China (Grant No. 81972160 by WP), and the Beijing Natural Science Foundation (Grant No. 17L20019 by WP). The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript." Please include your amended statements within your cover letter; we will change the online submission form on your behalf. 4. In your Data Availability statement, you have not specified where the minimal data set underlying the results described in your manuscript can be found. PLOS defines a study's minimal data set as the underlying data used to reach the conclusions drawn in the manuscript and any additional data required to replicate the reported study findings in their entirety. All PLOS journals require that the minimal data set be made fully available. For more information about our data policy, please see http://journals.plos.org/plosone/s/data-availability. Upon re-submitting your revised manuscript, please upload your study’s minimal underlying data set as either Supporting Information files or to a stable, public repository and include the relevant URLs, DOIs, or accession numbers within your revised cover letter. For a list of acceptable repositories, please see http://journals.plos.org/plosone/s/data-availability#loc-recommended-repositories. Any potentially identifying patient information must be fully anonymized. Important: If there are ethical or legal restrictions to sharing your data publicly, please explain these restrictions in detail. Please see our guidelines for more information on what we consider unacceptable restrictions to publicly sharing data: http://journals.plos.org/plosone/s/data-availability#loc-unacceptable-data-access-restrictions. Note that it is not acceptable for the authors to be the sole named individuals responsible for ensuring data access. We will update your Data Availability statement to reflect the information you provide in your cover letter. 5. Please include your full ethics statement in the ‘Methods’ section of your manuscript file. In your statement, please include the full name of the IRB or ethics committee who approved or waived your study, as well as whether or not you obtained informed written or verbal consent. If consent was waived for your study, please include this information in your statement as well. 6. Please include a copy of Table S8, S9, S2 and S4 which you refer to in your text on pages 6, 11 and 12. 7. Please include captions for your Supporting Information files at the end of your manuscript, and update any in-text citations to match accordingly. Please see our Supporting Information guidelines for more information: http://journals.plos.org/plosone/s/supporting-information. We prepared the new version on the basis of “Journal Requirements” mentioned in the Decision letter: 1. We referred to the The PLOS ONE style templates. 2. The correct grant numbers for the Funding are those in the Acknowledgments Section of the first version. That is, "This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700), the National Natural Science Foundation of China (Grant No. 81972160), and the Beijing Natural Science Foundation (Grant No. 17L20019)." 3. We have removed any funding-related text from the manuscript. Please move the following contents to the Funding Statement: "This research was funded by the National Key R&D Program of China (Grant Nos. 2018YFC2001400 and 2018YFC2001700), the National Natural Science Foundation of China (Grant No. 81972160), and the Beijing Natural Science Foundation (Grant No. 17L20019)." 4. The minimal data set has been included in the Supporting information. 5. The information on IRB has been added. 6. Table S8, S9, S2 and S4 should be Table 8, 9, 2 and 4. We have corrected the error in the revised version. 7. We have included captions for the Supporting Information files at the end of the manuscript, and updated in-text citations. ========================= Reviewer #1: This is an interesting article. There is real value. However, I have several concerns regarding details missing, and technical questions that are unanswered in the article in its present form. The article is generally well written, but all acronyms are not fully spelled out on first use, and software packages and links are not always provided in the methods. The authors would have made both review and their revision tasks easier if they had added line numbers in the draft. Response:We really appreciate the comments and suggestions. The revised manuscript, we hope, can solve the raised issues and is clearly expressed. I will detail the various data and methodology concerns in detail below. 1. ACRONYMS must be spelled out on first use and links to software and citations provided. The omissions are numerous. A few (not complete list) MDL, AMUSE, JADE, ERICA and surprisingly, even ICASSO which is a pivotal method here. Response: We did not find the origin of the acronym “ICASSO” when we prepare the manuscript. The developers provided a link http://research.ics.aalto.fi/ica/icasso/ where two publications (http://research.ics.aalto.fi/ica/icasso/publications.shtml) can be found. However, neither the content on the website nor the publications introduced the origin of ICASSO. We also failed to find the origin of the acronym in other publications. To find the answer, we asked the developer Prof. Aapo Hyvärinen and learn: it is not a real acronym. It is a kind of a joke on the acronym "ICA" and the name of the artist Picasso. We think this information is useful and have supplemented it in the Acknowledgments since no answer can be found publicly elsewhere. The origins of other acronyms, links to software, and citations have been added. 2. Only two human subject data sets are used in this paper. This might be acceptable if we were convinced that the data was sufficiently rich, but this information is missing. The reader needs to know fully data set sizes: numbers of imaging sessions, full number of tasks, full number of stimuli in tasks etc. The richness of data is a factor in ICA outcomes so these aspects MUST be reported. Similarly, the method of data setup for ICA needs better description. Was data analyzed in each subject ultimately as a single concatenated ensemble (and if so, of what matrix dimensions), or as a number of subset ensembles (and if so, of what dimensions, and how many such ensembles). Because only two individuals data were used, there remains the unfortunate possibility that there are structural individual-based idiosyncracies in the data here that favor Infomax. I do not believe this likely, but it cannot be fully discounted given the data. Other research applying ICA or testing dimensionality separation methods including ICA have used larger numbers of individuals, or/and artificial data generation under constraints to explore comparisons, or reliability. Examples in motor field are Tresch, D'Avella, Cheung, papers in J Neurophysiology for synergy separation analyses, and Yang, Logan and Giszter in PNAS exploring SCC-like measures on motor synergy outcomes across individual animals. An aspect not explored by the authors of this paper (and given lack of detail provided, this may or may not be possible in their data), that might be used to enrich the analysis they perform is bootstrapping or jack-knifing the data sets used, in addition to subsequent multiple iterations of ICA methods on each subjects subset data. At the very least, there needs to be discussion of the data limits in this paper resulting from only 2 subjects and only 1 subject per type of experiment, and the caveats resulting, i.e., possibility of individual idiosyncracy of data favoring specific ICA methods. Response: The fMRI data sets contained three types of status, sensory stimulation, imagined movement and motor execution task. The imagined movement is actually one type of cognitive task. We used these data to represent several different types. Based on the suggestion, missing information such as numbers of imaging sessions has been added. More details of the method of data setup for ICA have been added. Except the algorithm and the number of ICs, default settings/parameters defined by the GIFT software were used during analysis. There was a single ensemble for each subject since each data set contained only one subject. During data reduction steps, for one subject one session, the data reduction actually would be disabled since the number of principal components extracted from the data is the same as the number of independent components, as introduced by the manual of the GIFT software, i.e., the matrix dimensions would not be changed. (In group analysis containing a number of subjects, subjects in a group can also be concatenated as a single ensemble by the GIFT software). As you pointed out, some published studies such as Tresch, D'Avella, Cheung’s and Yang, Logan and Giszter’s have successfully performed ICA with Infomax. We suppose that many study groups have found some merits of Infomax empirically. The merit of this study is, on the aspect of reliability, to show the priority of Infomax and to uncover in which index Infomax is superior to other tested ICA algorithms in the single-subject level. We used the RandInit mode (algorithm starts with Randomizing different Initial values) to run ICASSO. This information has been added in the manuscript. The RandInit mode was chosen because 1) the RandInit mode in ICASSO uses the original data whereas the data will be resampled in the bootstrapping method; 2) the RandInit mode generates correlation coefficients with straightforward calculations whereas some extra normalization is necessary for bootstrapping [J. Himberg, A. Hyvärinen and F. Esposito. NeuroImage 2004(3):1214-1222.]. If we run ICASSO 10 times, the algorithm (e.g., Infomax) will run 10 times; in each time, the algorithm starts with randomizing different initial conditions. This study used fMRI data from two subjects. Discussions of this limitation have been added at the end of the manuscript. These fMRI data sets contained three types of status, i.e., sensory stimulation, imagined movements, and motor execution task. Single-subject ICA analysis is the foundation of group-level analysis. However, results from two single-subject analyses may not well represent the performance of an ICA algorithm for all types of data. Whether the reliability of Infomax is superior to other algorithms needs to be further proved with other types of data. 3. The SCC method as described in section 2.4 is unclear. It is possible to do this in two ways - correlate the individual IC spatial components picking best correlations (where pairings of and IC with other sets may not be unique- same IC best in two or more correlations), or correlate the spatial matrix (e.g., in MATLAB using matperm, matcorr) and then IC correlations will be unique, based on the matrix permutation use in the unique IC matchings. The authors used the former, non-unique method I believe, but this might have biased results. If they did, with the possibility of non-unique matching of ICs, then I think the reader needs to know if there were non-unique best IC spatial correlations in each method, and if so how many. This might be an additional metric on ICA algorithm quality. Response: The SCC was the correlation coefficient value between two IC spatial maps, that is, two spatial matrices. When comparing two groups of ICs A and B, correlation coefficient value between each IC in group A and each IC in group B was calculated. The best matched ICs must be a pair showing the highest correlation value. This best- matched pair is unique since the corresponding r value is the highest one; we did not find one IC in one group with two equal highest r values with two ICs in the other group. In our data, we only found lower thresholds: 1) (section 3.4 Difference in SCC values between the most reliable results and the other results for each non-deterministic algorithm) For the sensory data, if the SCC values ≤ 0.88 (found in the results of COMBI), there would be an IC (in the corresponding ICASSO result) that matched two IC maps of the most reliable result (i.e., presenting similar SCC values), and there was another IC (in the corresponding ICASSO result) that did not match any IC in the most reliable result. In other words, when compared with the most reliable result, the other nine results with a SCC ≤ 0.88 indicated unreliable performance; 2) (section 3.5 Comparing SCC values among the nine ICA algorithms). If there was an SCC value less than 0.669816, there would be at least one “unmatched” IC in the result of such algorithm, which means that this IC could not exclusively match any IC map in the results of Infomax and thus suggests a poor spatial consistency between the results of the algorithm and Infomax. This information has been presented in the manuscript. 4. Finally, in relation to the SCC statistics used: Correlation, especially here, is non-gaussian if used directly. This may present difficulties in interpreting the statistics in tables 6 and 7 that could be avoided. Using the Fisher-Z transform (see Yang, Logan, Giszter, PNAS 2019 for example with Infomax ICA data stats) the correlations are transformed to normal variables, improving the interpretation of parametric methods. Response: We used correlation coefficient value to compare different conditions. The correlation values can be transformed to normal variables with the Fisher-Z transform. The population correlation coefficient is ρ, and the sample correlation coefficient is r. The sample is a part of the population. When Fisher-Z the transformation is applied to the sample correlation coefficient r, the sampling distribution of the transformed variable is approximately normal. Without the Fisher transformation, the variance of r (between two variables X and Y) grows smaller as |ρ| gets closer to 1 and thus is not normally distributed. (referring to publications listed on https://en.wikipedia.org/wiki/Fisher_transformation) In our case, however, for each comparison between two groups of SCC, normal distribution of transformed SCC values in each group cannot be assured. The reason is that the SCC values (r values) come from different populations: For example, a SCC value between j-th IC in group A and k-th IC in group B can be Fisher-Z transformed, which leads to a normal-distributed transformed SCC value. Such normal distribution means the transformed r values meet normal distribution for the two populations where two samples (j-th IC in group A and k-th IC in group B) are drawn. But now we have other SCC values between other pairs of ICs (e.g., m-th IC in group A and n-th IC in group B). These are samples from other populations. Therefore, the transformed r values may not be normally distributed when putted into a group. For a large data set it may be possible (normally distributed), but we have only a small sample size here (12 ICs for the sensory data and 8 ICs for motor data). As the solution, we used the Kruskal–Wallis test in Table 6 and 7. The Kruskal–Wallis test is a nonparametric test that does not rely on normal distribution. 5. Equations would be helpful throughout the methods if very clearly written. The goal is reproducibility of methods. Response: We supplemented an equation in section 2.3: In ICASSO, the similarity between one pair of ICs (i and j) is quantified by the absolute value of their mutual correlation coefficients �  i j. The clustering process is performed by using distance between the two ICs. The distance is determined by transforming the similarity matrix into a dissimilarity (distance) matrix: di j=1-�  i j. There are two commonly used methods to transform the similarity matrix into a distance matrix: di j=1-�  i j or di j=1/�  i j. Therefore, the supplemented information is helpful if someone want to verify the results since the two methods generate different values. There are some long and complex equations in the cited literature. We do not present these equations in that they need detailed introductions and may interfere with the topic of the manuscript. 6. The data sharing statement is insufficient, for this study an anonymized repository should be chosen - there are many. Response: We have added the fMRI data to the Supporting information so that other groups can verify the results. This paper is a solid contribution, but marred by the omissions in detail, and possibly improved by attention to the technical points noted. Response: Thanks very much again for the comments and suggestions. We hope the new version responds well to the concerns. Submitted filename: Response to Reviewers.docx Click here for additional data file. 16 Mar 2022
PONE-D-21-19322R1
Comparing the reliability of different ICA algorithms for fMRI analysis
PLOS ONE Dear Dr. Wei, Thank you for submitting your manuscript to PLOS ONE. After careful consideration, we feel that it has merit but does not fully meet PLOS ONE’s publication criteria as it currently stands. Therefore, we invite you to submit a revised version of the manuscript that addresses the points raised during the review process. Please submit your revised manuscript by Apr 30 2022 11:59PM. If you will need more time than this to complete your revisions, please reply to this message or contact the journal office at plosone@plos.org. When you're ready to submit your revision, log on to https://www.editorialmanager.com/pone/ and select the 'Submissions Needing Revision' folder to locate your manuscript file. Please include the following items when submitting your revised manuscript:
If you would like to make changes to your financial disclosure, please include your updated statement in your cover letter. Guidelines for resubmitting your figure files are available below the reviewer comments at the end of this letter. A rebuttal letter that responds to each point raised by the academic editor and reviewer(s). You should upload this letter as a separate file labeled 'Response to Reviewers'. A marked-up copy of your manuscript that highlights changes made to the original version. You should upload this as a separate file labeled 'Revised Manuscript with Track Changes'. An unmarked version of your revised paper without tracked changes. You should upload this as a separate file labeled 'Manuscript'. If applicable, we recommend that you deposit your laboratory protocols in protocols.io to enhance the reproducibility of your results. Protocols.io assigns your protocol its own identifier (DOI) so that it can be cited independently in the future. For instructions see: https://journals.plos.org/plosone/s/submission-guidelines#loc-laboratory-protocols. Additionally, PLOS ONE offers an option for publishing peer-reviewed Lab Protocol articles, which describe protocols hosted on protocols.io. Read more information on sharing protocols at https://plos.org/protocols?utm_medium=editorial-email&utm_source=authorletters&utm_campaign=protocols. We look forward to receiving your revised manuscript. Kind regards, Kewei Chen, Ph.D Academic Editor PLOS ONE Journal Requirements: Please review your reference list to ensure that it is complete and correct. If you have cited papers that have been retracted, please include the rationale for doing so in the manuscript text, or remove these references and replace them with relevant current references. Any changes to the reference list should be mentioned in the rebuttal letter that accompanies your revised manuscript. If you need to cite a retracted article, indicate the article’s retracted status in the References list and also include a citation and full reference for the retraction notice. [Note: HTML markup is below. Please do not edit.] Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: All comments have been addressed ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: No ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: This is a nice revision. Thank you for the work on the term ICASSO origin. This revision is coupled with a very responsive letter, but not all comments in the letter appear in the revised manuscript as I read it. Since readers may have the same concerns as reviewers it is crucial to include these in the manuscript, unless the full review history is also going to be published. In particular, the issue I could not find in the materials and methods: That the component correlation method used (in section "Comparing the reliability among four ICA algorithms with SCC") did not in the present data produce any double use correlations must be stated, and that this was observed to be true, although this is not guaranteed by the method. I still wasn't clear if the authors were relying here on other code or directly checked this in their work, and this should be clearly stated. Line 331 - probably, term 'well-performing' is better than 'well-performed' here ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No [NOTE: If reviewer comments were submitted as an attachment file, they will be attached to this email and accessible via the submission site. Please log into your account, locate the manuscript record, and check for the action link "View Attachments". If this link does not appear, there are no attachment files.] While revising your submission, please upload your figure files to the Preflight Analysis and Conversion Engine (PACE) digital diagnostic tool, https://pacev2.apexcovantage.com/. PACE helps ensure that figures meet PLOS requirements. To use PACE, you must first register as a user. Registration is free. Then, login and navigate to the UPLOAD tab, where you will find detailed instructions on how to use the tool. If you encounter any issues or have any questions when using PACE, please email PLOS at figures@plos.org. Please note that Supporting Information files do not need this step.
28 Mar 2022 Thank you very much for the comments and suggestions. The explanations in the former letter have been added in the manuscript if the contents were not included in the former version. In section "Comparing the reliability among four non-deterministic ICA algorithms with SCC” (and other sections where SCCs were calculated), the SCC was calculated by using MATLAB function corrcoef. We have added this information in the manuscript. In this section, the SCC values are used in two steps. Step 1 is a simple picking-the maximum-value process, introduced in the second paragraph of this section. This is not a statistical comparison but just picks the highest SCC value. As a result, we get a list of SCC values for each pair of groups. Step 2 is a statistical comparison between different lists of SCCs by using the Kruskal–Wallis test, introduced in the last paragraph of this section. Thus, this is not a double use of correlations. Step 1 provides values for statistical analysis in Step 2. In the new version, we have added this information and highlighted supplemented words. We hope the contents will be more clearly expressed. The word 'well-performed' has been changed to 'well-performing' based on your suggestion. Submitted filename: Response to Reviewers.docx Click here for additional data file. 14 Jun 2022 Comparing the reliability of different ICA algorithms for fMRI analysis PONE-D-21-19322R2 Dear Dr. Wei, We’re pleased to inform you that your manuscript has been judged scientifically suitable for publication and will be formally accepted for publication once it meets all outstanding technical requirements. Within one week, you’ll receive an e-mail detailing the required amendments. When these have been addressed, you’ll receive a formal acceptance letter and your manuscript will be scheduled for publication. An invoice for payment will follow shortly after the formal acceptance. To ensure an efficient process, please log into Editorial Manager at http://www.editorialmanager.com/pone/, click the 'Update My Information' link at the top of the page, and double check that your user information is up-to-date. If you have any billing related questions, please contact our Author Billing department directly at authorbilling@plos.org. If your institution or institutions have a press office, please notify them about your upcoming paper to help maximize its impact. If they’ll be preparing press materials, please inform our press team as soon as possible -- no later than 48 hours after receiving the formal acceptance. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information, please contact onepress@plos.org. Kind regards, Pew-Thian Yap Academic Editor PLOS ONE Additional Editor Comments (optional): Reviewers' comments: Reviewer's Responses to Questions Comments to the Author 1. If the authors have adequately addressed your comments raised in a previous round of review and you feel that this manuscript is now acceptable for publication, you may indicate that here to bypass the “Comments to the Author” section, enter your conflict of interest statement in the “Confidential to Editor” section, and submit your "Accept" recommendation. Reviewer #1: (No Response) ********** 2. Is the manuscript technically sound, and do the data support the conclusions? The manuscript must describe a technically sound piece of scientific research with data that supports the conclusions. Experiments must have been conducted rigorously, with appropriate controls, replication, and sample sizes. The conclusions must be drawn appropriately based on the data presented. Reviewer #1: Yes ********** 3. Has the statistical analysis been performed appropriately and rigorously? Reviewer #1: Yes ********** 4. Have the authors made all data underlying the findings in their manuscript fully available? The PLOS Data policy requires authors to make all data underlying the findings described in their manuscript fully available without restriction, with rare exception (please refer to the Data Availability Statement in the manuscript PDF file). The data should be provided as part of the manuscript or its supporting information, or deposited to a public repository. For example, in addition to summary statistics, the data points behind means, medians and variance measures should be available. If there are restrictions on publicly sharing data—e.g. participant privacy or use of data from a third party—those must be specified. Reviewer #1: Yes ********** 5. Is the manuscript presented in an intelligible fashion and written in standard English? PLOS ONE does not copyedit accepted manuscripts, so the language in submitted articles must be clear, correct, and unambiguous. Any typographical or grammatical errors should be corrected at revision, so please note any specific errors here. Reviewer #1: No ********** 6. Review Comments to the Author Please use the space provided to explain your answers to the questions above. You may also include additional comments for the author, including concerns about dual publication, research ethics, or publication ethics. (Please upload your review as an attachment if it exceeds 20,000 characters) Reviewer #1: I think the use of Fisher Z transforms on your data would be necessary for parametric tests, and your argument about the distributions of i th and jth correlation pairings potentially having different distributions is not particularly good, and only meaningful insofar as your number of subjects is too small. The correlations have different normal distributions, as you correctly state, but the central limit theorem indicates a sum of normal distributions will tend to normal. The distribution problems you may have with normality in your study are primarily due to the very small subject sample size you have (2 subjects) meaning the distribution might be multimodal if variances are low for different correlations across your analysis runs, because not enough subject variations for distributions are contributing. It would be better to directly acknowledge this as the basis of the non parametric testing. Since you are choosing using non-parametric tests here, rather than parametric tests, the ordinal testing you used will be unaffected by the z transform and it is not needed. The best study would have 6-10 subjects per group and use z-transformed correlation data. I suggest a note to this effect, but it is your choice. ********** 7. PLOS authors have the option to publish the peer review history of their article (what does this mean?). If published, this will include your full peer review and any attached files. If you choose “no”, your identity will remain anonymous but your review may still be made public. Do you want your identity to be public for this peer review? For information about this choice, including consent withdrawal, please see our Privacy Policy. Reviewer #1: No ********** 17 Jun 2022 PONE-D-21-19322R2 Comparing the reliability of different ICA algorithms for fMRI analysis Dear Dr. Wei: I'm pleased to inform you that your manuscript has been deemed suitable for publication in PLOS ONE. Congratulations! Your manuscript is now with our production department. If your institution or institutions have a press office, please let them know about your upcoming paper now to help maximize its impact. If they'll be preparing press materials, please inform our press team within the next 48 hours. Your manuscript will remain under strict press embargo until 2 pm Eastern Time on the date of publication. For more information please contact onepress@plos.org. If we can help with anything else, please email us at plosone@plos.org. Thank you for submitting your work to PLOS ONE and supporting open access. Kind regards, PLOS ONE Editorial Office Staff on behalf of Dr. Pew-Thian Yap Academic Editor PLOS ONE
Table 1

Repetition number k did not influence Iq values for non-deterministic ICA (sensory data).

InfomaxFastICAEVDCOMBI
Chi-sq1.041.1111.670.8406
P0.99930.99920.23230.9744

For each non-deterministic algorithm, the Iq values of different repetition number k (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100) did not present statistically significant difference (Kruskal–Wallis test). The result of the Kruskal–Wallis test of COMBI was based on six groups of Iq values because four results (when k = 10, 20, 40, and 70) of COMBI had clusters without the Iq value.

Table 3

Repetition number k did not influence Iq values for non-deterministic ICA (motor data).

InfomaxFastICAEVDCOMBI
Chi-sq15.378.831.999.14
P0.08140.45350.99160.4241

For each non-deterministic algorithm, the Iq values of different repetition number k (k = 10, 20, 30, 40, 50, 60, 70, 80, 90, and 100) did not present statistically significant difference (Kruskal–Wallis test).

Table 8

Range of SCC values between the most reliable ICASSO results and the other nine results (sensory data).

InfomaxFastICAEVDCOMBI
Max10.9999990.9993571.000000
Min0.9693510.0054710.9948960.441637
Number of clusters with SCC<0.90507
Table 9

Range of SCC values between the most reliable ICASSO results and the other nine results (motor data).

InfomaxFastICAEVDCOMBI
Max1111
Min0.9982380.9976700.9987310.998982
Number of clusters with SCC<0.90000
Table 10

SCC values between the most reliable Infomax results and the other nine results (sensory data).

AMUSEERICAJADERADICALSIMBECFastICAEVDCOMBI
Median0.7080410.6560050.955730.9109220.6857520.99729180.517670.991084
Max0.9166480.9921570.9982840.9966360.9925120.99992350.8194380.999003
Min0.5199980.2799980.0962910.0337010.1869710.97861250.2442010.018344

The results of FastICA (the most reliable results) presented higher SCC values than other algorithms.

Table 11

SCC values between the most reliable Infomax results and the other nine results (motor data).

AMUSEERICAJADERADICALSIMBECFastICAEVDCOMBI
Median0.8014760.750610.941870.987480.7393440.9988790.6235380.99123
Max0.9770520.9392870.9997950.9983790.9927990.99989370.975870.999588
Min0.6528540.4762540.6698160.7236420.4917130.9633210.2431150.88953
  17 in total

1.  Analysis of fMRI data by blind separation of data in a tiny spatial domain into independent temporal component.

Authors:  Huafu Chen; Dezhong Yao; Yan Zhuo; Lin Chen
Journal:  Brain Topogr       Date:  2003       Impact factor: 3.020

2.  Validating the independent components of neuroimaging time series via clustering and visualization.

Authors:  Johan Himberg; Aapo Hyvärinen; Fabrizio Esposito
Journal:  Neuroimage       Date:  2004-07       Impact factor: 6.556

3.  Ranking and averaging independent component analysis by reproducibility (RAICAR).

Authors:  Zhi Yang; Stephen LaConte; Xuchu Weng; Xiaoping Hu
Journal:  Hum Brain Mapp       Date:  2008-06       Impact factor: 5.038

4.  Stability of ICA decomposition across within-subject EEG datasets.

Authors:  Romain Grandchamp; Claire Braboszcz; Scott Makeig; Arnaud Delorme
Journal:  Annu Int Conf IEEE Eng Med Biol Soc       Date:  2012

5.  An evaluation of independent component analyses with an application to resting-state fMRI.

Authors:  Benjamin B Risk; David S Matteson; David Ruppert; Ani Eloyan; Brian S Caffo
Journal:  Biometrics       Date:  2013-12-18       Impact factor: 2.571

6.  Manual selection of spontaneous activity maps derived from independent component analysis: criteria and inter-rater reliability study.

Authors:  Daniel R Roquet; Bich-Tuy Pham; Jack R Foucher
Journal:  J Neurosci Methods       Date:  2013-12-04       Impact factor: 2.390

7.  Variability of ICA decomposition may impact EEG signals when used to remove eyeblink artifacts.

Authors:  Matthew B Pontifex; Kathryn L Gwizdala; Andrew C Parks; Martin Billinger; Clemens Brunner
Journal:  Psychophysiology       Date:  2016-12-27       Impact factor: 4.016

8.  A simple and objective method for reproducible resting state network (RSN) detection in fMRI.

Authors:  Gautam V Pendse; David Borsook; Lino Becerra
Journal:  PLoS One       Date:  2011-12-12       Impact factor: 3.240

9.  Performance of blind source separation algorithms for fMRI analysis using a group ICA method.

Authors:  Nicolle Correa; Tülay Adali; Vince D Calhoun
Journal:  Magn Reson Imaging       Date:  2006-12-08       Impact factor: 2.546

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.