Literature DB >> 26594340

Creating 3D visualizations of MRI data: A brief guide.

Christopher R Madan1.   

Abstract

While magnetic resonance imaging (MRI) data is itself 3D, it is often difficult to adequately present the results papers and slides in 3D. As a result, findings of MRI studies are often presented in 2D instead. A solution is to create figures that include perspective and can convey 3D information; such figures can sometimes be produced by standard functional magnetic resonance imaging (fMRI) analysis packages and related specialty programs. However, many options cannot provide functionality such as visualizing activation clusters that are both cortical and subcortical (i.e., a 3D glass brain), the production of several statistical maps with an identical perspective in the 3D rendering, or animated renderings. Here I detail an approach for creating 3D visualizations of MRI data that satisfies all of these criteria. Though a 3D 'glass brain' rendering can sometimes be difficult to interpret, they are useful in showing a more overall representation of the results, whereas the traditional slices show a more local view. Combined, presenting both 2D and 3D representations of MR images can provide a more comprehensive view of the study's findings.

Entities:  

Keywords:  MRI visualization; ROI; brain mapping; fMRI; image processing; neuroanatomy

Year:  2015        PMID: 26594340      PMCID: PMC4648228          DOI: 10.12688/f1000research.6838.1

Source DB:  PubMed          Journal:  F1000Res        ISSN: 2046-1402


Introduction

When presenting and publishing findings of magnetic resonance imaging (MRI) studies, sometimes it is difficult to adequately present the results because they are 3D, while papers and slides can inherently only be 2D. A solution is to create figures that include perspective and can convey 3D information, and the creation of such figures can be readily produced using standard functional magnetic resonance imaging (fMRI) analysis packages, such as SPM ( http://www.fil.ion.ucl.ac.uk/spm/), AFNI ( http://afni.nimh.nih.gov/afni/; with SUMA), and FreeSurfer ( https://surfer.nmr.mgh.harvard.edu), as well as some more specialty programs, such as MRIcroGL ( http://www.mccauslandcenter.sc.edu/mricrogl/), 3DSlicer ( http://www.slicer.org), and Mango ( http://ric.uthscsa.edu/mango/). While these numerous options can provide 3D renderings of MRI data, many of them are unable to provide useful functionality such as visualizing activation clusters that are both cortical and subcortical, i.e., a 3D glass brain. This difficulty increases further if one wants to produce 3D renderings of several activation maps with an identical perspective (e.g., camera angle) or animated renderings (e.g., a rotating 3D glass brain). Here I briefly detail a straightforward approach for creating 3D visualizations of MRI data that work in these scenarios, as well as readily generalize to most other instances. An illustration of this processing workflow is shown in Figure 1. An additional example of making a 3D rendering of traced regions of interest (ROIs) is also outlined.
Figure 1.

Overview of the processing pipeline.

The guide will primarily utilize two programs, ITK-SNAP (v. 3.0.0; Figure 2A; http://www.itksnap.org; Yushkevich ) and ParaView (v. 4.3.1; Figure 2B; http://www.paraview.org; Ayachit, 2015). Both programs are available for both Windows and Mac operating systems and are freely available. Data files produced in the examples are provided in the Supplementary material (see Appendix A).
Figure 2.

Screenshots of ( a) ITK-SNAP and ( b) ParaView.

Screenshots of ( a) ITK-SNAP and ( b) ParaView.

Methods

Procedure #1: Visualizing cluster maps in a glass brain

As a first exercise in visualizing MRI data in 3D, we will start with a statistical map. Depending on where your maps are coming from, you may need to apply a height threshold (i.e., t- or Z-critical) and/or a minimum cluster extend threshold ( k). As a starting point and to make this guide more general and more reproducible, I will start with a statistical map obtained from NeuroSynth ( http://www.neurosynth.org; Yarkoni ), which will be in NIfTI format (Neuroimaging Informatics Technology Initiative; http://nifti.nimh.nih.gov). Briefly, NeuroSynth conducts automated meta-analyses across thousands of fMRI studies by calculating a frequency metric for how often specific terms are mentioned in the paper (e.g., “memory”, “emotion”) in relation to voxels reported in the results tables. See Yarkoni for further details. As an example of how to obtain thresholded statistical maps from SPM analyses, see Appendix B. For this example I searched the online version of NeuroSynth for the term “memory” ( http://www.neurosynth.org/analyses/terms/memory/; see Figure 3A). I used the “forward inference” map as the example statistical map, along with the anatomical volume provided (obtained by clicking the download buttons displayed to the right of the layer names). For these examples, un-gzip the NIfTI volumes from NeuroSynth. Rename the forward inference map file from memory_pAgF_z_FDR_0̷.0̷1.nii to statmap.nii.
Figure 3.

Images of the statistical map used in Procedure #1.

( a) Obtaining the statistical map from NeuroSynth.org. ( b) Coronal slices of the thresholded activation clusters. ( c) 3D renderings of the clusters from two different perspectives. ( d) Stereoscopic anaglyph 3D rendering of the first perspective shown in panel C, to be viewed using red-blue 3D glasses.

Images of the statistical map used in Procedure #1.

( a) Obtaining the statistical map from NeuroSynth.org. ( b) Coronal slices of the thresholded activation clusters. ( c) 3D renderings of the clusters from two different perspectives. ( d) Stereoscopic anaglyph 3D rendering of the first perspective shown in panel C, to be viewed using red-blue 3D glasses. Since the anatomical 3D surface meshes are generally usable, in addition to outlining the steps for creating this glass brain volume, the resulting surface mesh file is also provided as Supplementary material (see Appendix A). While there is an abundance of anatomical volumes in normalized template space, here we will use the one provided on NeuroSynth (click “anatomical” where shown in Figure 2A). Before the map can be rendered in 3D, both the height and cluster-extent thresholds should be applied. In some fMRI analysis packages this can be output directly (e.g., see Appendix B). If this is the case, export the threshold cluster image and skip to section 1.3; if this is not the case, we will manually apply these thresholds ourselves. Here I will use examples of how to manually apply these thresholds using MATLAB (R2013a; The MathWorks Inc., Natick, MA) and SPM8 ( http://www.fil.ion.ucl.ac.uk/spm/; Wellcome Department of Cognitive Neurology, UCL, London, UK), though other packages are able to do this as well. (See Madan, 2014, for an introductory guide to MATLAB.) Manually applying the height and cluster-extent threshold is a bit cumbersome. Using the imcalc function in SPM, we can easily apply a height threshold to our volume by outputting a binary volume, where the voxel intensity statistic (i.e., from a t-, F-, or Z-statistic map) is above the threshold. Since our NeuroSynth memory image has a large number of highly significant clusters, we will threshold our statmap.nii to isolate the voxels where the statistic ( Z-value) is above 12 (MATLAB code shown below). In the current case, lower thresholds yielded large clusters, which made the figure less interpretable (i.e., many regions comprised a single cluster, making it difficult to view the topology of the regions from the 3D view). When plotting results from your own fMRI study, you would likely use a threshold around 3 for the t- or Z-statistical map (corresponding to approximately p<.001). To apply the cluster-extent threshold, we will use the nii_threshreslicecluster function (freely available from http://www.mccauslandcenter.sc.edu/CRNL/tools/spm8-scripts) to isolate clusters of voxels of at least a minimum volume of 400 mm 3. Again, this value can be adjusted, and usually would be set higher than you would use for your statistical analyses, as the 3D rendering is intended more to provide a global view of the significant clusters, and is encumbered by the inclusion of many small clusters. A volume of 400 mm 3 corresponds to 50 voxels where the voxel size is 2 mm-isotropic. The function can also apply height thresholds, but it thresholds rather than binarizes the image (i.e., converting it to a mask), which is not as useful for our current purposes. The output file from this command will automatically be named rstatmapH.nii, rename it to statmapThresh.nii. The height threshold should also be applied to the anatomical volume, anatomical.nii, but the cluster threshold is unnecessary. The resulting output file will be named anatomicalH.nii, rename it to glassbrain.nii. To visualize the NIfTI volumes in 3D, we need to convert the voxel data into a 3D surface mesh in the VTK (Visualization ToolKit) format. Designed for anatomical tracing, ITK-SNAP includes this functionality. The simplest way to do so is to load each volume as both the main volume and as the segmentation volume. If you use the structural volume as the main volume and the statistical map as the segmentation, you may have issues with the bounding boxes not matching. Since we will move to another program with our 3D surfaces, it does not matter if the bounding boxes match or not. Make sure that the volume is loaded correctly, as shown in Figure 2A. While ITK-SNAP can render 3D volumes, as shown in the bottom left portion of the screenshot shown in Figure 2A, its rendering options are limited. For instance, if you want to render several volumes in 3D from a consistent perspective/camera angle, ITK-SNAP is unable to accommodate; while ITK-SNAP can temporarily store camera information, this perspective information is lost if the program is closed or crashes, and it cannot be saved for later use nor can it be manually specified. As a result, it will be impossible to obtain the exact same camera angle. To rectify this shortcoming, we will make our 3D renderings in ParaView, which also has additional useful features. To export the meshes in VTK format from ITK-SNAP, use the menus to navigate to Segmentation, then “Export as Surface Mesh…”. Next, choose “Export meshes for all labels as a single scene” and save the file as a “VTK PolyData File”. In the current example there is only one surface mesh in each volume, but this is not always the case, such as in the ROI example discussed later. Note, it is possible to export volume data, rather than surface mesh data, as a VTK file in ITK-SNAP, but these files will not work with ParaView in the next step. If your VTK file does not work, double check that it was correctly exported as a surface mesh. Repeat these steps for both the statistical map and anatomical volume. Start ParaView and open your two new VTK files within the same scene. ParaView can be a bit overwhelming at first, but it has many useful features for rendering and setting up the camera. With some adjustment of the colors and opacity for the two surfaces, it should be fairly easy to produce a set up in ParaView similar to Figure 2B. You can rotate the camera manually using the mouse, and can reset the camera position with the buttons labelled “+X” through “−Z”. When the scene state is saved, the camera position is preserved in the scene file, allowing you to easily load another statistical map at a later time. The scene as a whole can be saved by selecting “File” then “Save State…” (PVSM format). The final renderings produced here are shown in Figure 3C, corresponding to the series of coronal slices shown in Figure 3B. Renderings can be saved using either “File” then “Save Screenshot…” or “Export Scene…”. Screenshots will always be exported as raster (i.e., pixel) images, while ‘exported scenes’ are vector/polygon based. Note that exported PDFs can also be based on “rasterize 3D geometry” (there is a checkbox). If you are unsure what you require, a screenshot is likely sufficient, but do try and experiment to find out what settings best meet your needs, as this overview of ParaView’s functionality is far from comprehensive. ParaView can also render stereoscopic 3D figures (e.g., anaglyph [red-blue], side-by-side) with a variety of 3D-compatible glasses options. An example of a red-blue stereoscopic render is shown in Figure 3D.

Procedure #2: Visualizing anatomical ROIs

For this example, I extracted several regions of the medial temporal cortex (hippocampus, amygdala, parahippocampal gyrus, fusiform gyrus) from the right hemisphere of the Hammers maximum probability atlas (n30r83; http://biomedic.doc.ic.ac.uk/brain-development/index.php?n=Main.AdultMaxProb). Regions were extracted using the imcalc tool included in SPM8, such that each ROI corresponded to a unique intensity value (1=hippocampus, 2=amygdala, 3=parahippocampal gyrus, 4=fusiform gyrus): The ROI is shown plotted over a structural volume in Figure 4A.
Figure 4.

Images of the anatomical ROIs used in Procedure #2.

( a) Coronal slices of the anatomical ROIs. Panels B-E depict different 3D rendering settings of the ROIs, ( b) points along the surface of the ROIs, ( c) wireframe, ( d) surface, and ( e) surface rendering with specular. Note that the structural image used in panel A is from a different source than the traced ROIs, and thus they do not perfectly align.

Images of the anatomical ROIs used in Procedure #2.

( a) Coronal slices of the anatomical ROIs. Panels B-E depict different 3D rendering settings of the ROIs, ( b) points along the surface of the ROIs, ( c) wireframe, ( d) surface, and ( e) surface rendering with specular. Note that the structural image used in panel A is from a different source than the traced ROIs, and thus they do not perfectly align. As before, use ITK-SNAP to load the NIfTI volume and convert it to a VTK surface mesh. If you have multiple surfaces in the same volume, as we do here, be sure to select “Export meshes for all labels as a single scene” when exporting the surfaces. Start ParaView and load the VTK file, as done previously. As shown in Figure 4B–E, the volumes can be rendered as points, wireframes, and surfaces. Furthermore, many settings can be customized to adjust the rendering properties, such as the lighting/reflectance properties shown in Figure 4D and 4E. ParaView can also create cameras that move over time, allowing for the generation of animations of the structures rotating. This can be done using the “Animation View” panel in the bottom-center of ParaView: select “Camera”, “Orbit”, and then “+”. The default settings for the camera positions are usually sufficient. If desired, the camera path can also be edited afterwards by inputting specific coordinates (the best way to preview the path is to simply press ‘play’ at the top and see how it looks). Even without rendering the animation itself, having a camera path allows for later reproduction of 3D renderings from the same camera positions. Using a camera path, an animation can be rendered by going “File”, “Save Animation”. An example rendered video is shown in Movie 1. (Note, videos here were re-compressed with Handbrake [ https://www.handbrake.fr; freely available for Windows and Mac] to reduce their file size).

Movie 1. Rotating 3D animation of the anatomical ROI with the same render settings as used in Figure 4E.

Movie 1. Rotating 3D animation of the anatomical ROI with the same render settings as used in Figure 4E. Click here for additional data file.

Additional examples

Using the techniques discussed thus far, it is possible to create an image such as that shown in Figure 5, where the hippocampus is shown within a glass brain for a number of different species, using freely available brain atlases. Each panel was rendered separately, but all of the surface meshes were loaded into the same scene in ParaView. By additionally adding a plane with a checkerboard texture, it is also easy to present the scale of the structures. See Appendix C for details regarding each of the brain atlases.
Figure 5.

Comparative neuroanatomy of whole-brain and hippocampal brain volumes.

The square grid included in each panel measures 20 mm across, with each grid square subtending 2 mm. See Appendix C for details and references describing each brain atlas.

Comparative neuroanatomy of whole-brain and hippocampal brain volumes.

The square grid included in each panel measures 20 mm across, with each grid square subtending 2 mm. See Appendix C for details and references describing each brain atlas. With a few additional steps, more intricate 3D renderings can also be produced. For instance, if the anatomical volume is down-sampled while in NIfTI format, the resulting surface mesh is less dense and can be rendered as a wireframe, as shown in Figure 6A–C. For demonstration purposes, if the lower-resolution anatomical volume was subsequently up-sampled, the resulting high-density mesh is ‘blocky’ ( Figure 6D). If a combination of different densities of anatomical surface meshes are used together, e.g., the meshes from Figure 6A–C, along with a ROI, a rendering such as Figure 6E can be produced.
Figure 6.

Wireframes of brain images of different resolutions.

( a) Wireframe rendering of brain surface mesh produced from the Hammers atlas, which originally has voxel size of 1 mm-isotropic. Panels B- C show the wireframes of surfaces meshes made after first downsampling the volume to 5 or 10 mm-isotropic, respectively, resulting in less dense wireframes. ( d) For demonstration purposes, the result of upsampling the 10 mm-isotropic mesh (panel C) back to 1mm-isotropic. ( e) Rendering produced by combining the surface meshes used in panels A-C, along with an anatomical ROI of the hippocampus.

Wireframes of brain images of different resolutions.

( a) Wireframe rendering of brain surface mesh produced from the Hammers atlas, which originally has voxel size of 1 mm-isotropic. Panels B- C show the wireframes of surfaces meshes made after first downsampling the volume to 5 or 10 mm-isotropic, respectively, resulting in less dense wireframes. ( d) For demonstration purposes, the result of upsampling the 10 mm-isotropic mesh (panel C) back to 1mm-isotropic. ( e) Rendering produced by combining the surface meshes used in panels A-C, along with an anatomical ROI of the hippocampus. Figure 7 and Movie 2 show a few additional rendering examples from freely available data. fMRI activity related to finger tapping is shown in Figure 7A, with data obtained from Gorgolewski (2013; http://www.neurovault.org/collections/63/; full dataset available at: http://www.openfmri.org/dataset/ds000114). Striatal anatomy is shown in Figure 7B (Oxford-GSK-Imanova Structural Striatal Atlas, from FSL; Tziortzi ; http://fsl.fmrib.ox.ac.uk/fsl/fslwiki/Atlases/striatumstruc). DTI tractography showing 20 structures at varying levels of probability estimates is shown in Figure 7C and Movie 2 (JHU white-matter tractography atlas, from FSL; Hua ).
Figure 7.

Additional examples of 3D renderings of MRI data.

( a) Glass brain rendering of fMRI clusters associated with finger tapping, based on data from Gorgolewski , from three perspectives. ( b) Glass brain rendering of the anatomical ROIs included in the Oxford-GSK-Imanova Structural Striatal Atlas ( c) 3D rendering of the structures included in the JHU white-matter tractography atlas, with different mesh properties used for the 0%, 25%, and 50% probability estimates from the maximum probability volumes. See main text for additional details on the sources of the MRI data.

Additional examples of 3D renderings of MRI data.

( a) Glass brain rendering of fMRI clusters associated with finger tapping, based on data from Gorgolewski , from three perspectives. ( b) Glass brain rendering of the anatomical ROIs included in the Oxford-GSK-Imanova Structural Striatal Atlas ( c) 3D rendering of the structures included in the JHU white-matter tractography atlas, with different mesh properties used for the 0%, 25%, and 50% probability estimates from the maximum probability volumes. See main text for additional details on the sources of the MRI data.

Movie 2. Rotating 3D animation of the white-matter tractography atlas with the same render settings as used in Figure 8C.

Movie 2. Rotating 3D animation of the white-matter tractography atlas with the same render settings as used in Figure 8C. Click here for additional data file.

Conclusion

Though a 3D ‘glass brain’ rendering of fMRI activations can sometimes be difficult to interpret, they are useful in showing a more overall representation of which regions are activated, whereas the traditional slices show a more local view of the results. When the goal is to show anatomical structures, 3D figures are definitively more useful in conveying the 3D structure of the regions, as shown in the examples here. Combined, 2D and 3D representations of MR images can provide a more comprehensive view of the results, particularly when at least two 3D perspectives are shown, allowing for some ability to provide depth information.

Data availability

The data referenced by this article are under copyright with the following copyright statement: Copyright: © 2015 Madan CR Data associated with the article are available under the terms of the Creative Commons Zero "No rights reserved" data waiver (CC0 1.0 Public domain dedication). Figshare: Movie 1. Rotating 3D animation of the anatomical ROI with the same render settings as used in Figure 4E. doi: 10.6084/m9.figshare.1499152 ( Madan, 2015a). Figshare: Movie 2. Rotating 3D animation of the white-matter tractography atlas with the same render settings as used in Figure 8C. doi: 10.6084/m9.figshare.1499153 ( Madan, 2015b). This is a very useful guide to an important issue that is currently largely overlooked in the literature; producing high-quality presentations of brain imaging results that are informative, clear, and useful. The article is comprehensive and easy to follow, and the examples provided are appropriate, and produce very attractive images. This is an extremely useful paper that deserves wide readership in the field. While I agree with the author that ‘glass-brain’ visualisations are extremely useful for providing a comprehensive overview of patterns of brain activity in fMRI experiments, that doesn’t mean that conventional 2D slice views are not also useful. In fact, 2D views of particular activation clusters are really the only way to get a good idea of the precise position of a cluster, in relation to the sulcal/gyral anatomy, which is often important. An optimal strategy for comprehensive visualisation and localisation might then be to combine 2D and 3D views of results in the same figure. The author has done this more-or-less in Figure 3 (which includes coronal slices), but I wonder if perhaps an additional example figure which combines 2D and 3D views might be helpful? Perhaps as an example of the kinds of ‘real’ figures that could be produced for publications and presentations. Minor points of grammar, etc.: Abstract: "they are useful in showing a more overall representation of the results" More overall? Somewhat clumsy; replace with "more general" or just "overall". Page 2 first paragraph: "Here I briefly detail a straight- forward approach for creating 3D visualizations of MRI data that work in these scenarios, as well as readily generalize to most other instances." Something wrong with the tenses here; would suggest: "Here I briefly detail a straight- forward approach for creating 3D visualizations of MRI data that works in these scenarios, and also readily generalizes to most other instances." Page 4. Section on obtaining and thresholding the images. Fine, but the procedure outlined here is pretty cumbersome, as the author admits! This procedure might be optimal for those who use SPM as their primary analysis tool, but the 'fslmaths' function included with FSL could achieve this in a single command-line entry. Maybe include a sentence saying something like "Other options for thresholding are available, such as the basic functions included with FSL." I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. I think that this is a useful paper. Here are some minor comments You do not mention anything about multiple comparisons for the thresholding. I understand that these visualizations are mainly for obtaining a better understanding of the brain activation, but it would still be nice to mention the problems of multiple testing. For cluster level inference, I prefer if a cluster p-value threshold is used, and not an arbitrary cluster size like 400 mm 3 or 50 voxels. Cluster p-values can be obtained through parametric methods (Gaussian random field theory, available in SPM and FSL) or non-parametric methods (permutation testing, available in SnPM, FSL and BROCCOLI). I know that a very common approach is to use a cluster defining threshold of p = 0.001 or p = 0.005 (uncorrected for multiple comparisons), combined with an arbitrary cluster size threshold of 10 voxels. Such approaches should in my opinion be avoided, since the method is ad-hoc; it is impossible to know what the (corrected) p-value is for the combined procedure. The following paper may be of interest: Choong-Wan Woo, Anjali Krishnan, Tor D. Wager, Cluster-extent based thresholding in fMRI analyses: Pitfalls and recommendations, NeuroImage, Volume 91, 1 May 2014, Pages 412-419, ISSN 1053-8119, http://dx.doi.org/10.1016/j.neuroimage.2013.12.058 ------- You may mention two additional pieces of software, pysurfer and MevisLab. Pysurfer is a python tool for visualizing cortical surface representations https://pysurfer.github.io/ MevisLab is a free software that can be used for image processing and visualization. MevisLab includes functions from the libraries VTK and ITK, and it is easy to setup more advanced volume rendering pipelines, where you for example have several volume renderers, clip planes and more advanced transfer functions. http://www.mevislab.de/ ------- You do not mention anything about visualization research regarding fMRI. A more advanced way to visualize brain activation is to treat the activation as a light source in the anatomical volume, making the activity "glow" from the inside. You could include some of the following papers. Nguyen, T. K., Eklund, A., Ohlsson, H., Hernell, F., Ljung, P., Forsell, C., Andersson, M., Knutsson, H., Ynnerman, A., Concurrent Volume Visualization of Real-time fMRI, Proceedings of the 8th IEEE/EG International Conference on Volume Graphics, 53-60, 2010, http://dx.doi.org/10.2312/VG/VG10/053-060 Janoos, F., Nouanesengsy, B., Machiraju, R., Shen, H. W., Sammet, S., Knopp, M. and Mórocz, I. Á. (2009), Visual Analysis of Brain Activity from fMRI Data. Computer Graphics Forum, 28: 903–910. doi: 10.1111/j.1467-8659.2009.01458.x Jainek, W. M., Born, S., Bartz, D., Straßer, W. and Fischer, J. (2008), Illustrative Hybrid Visualization and Exploration of Anatomical and Functional Brain Data. Computer Graphics Forum, 27: 855–862. doi: 10.1111/j.1467-8659.2008.01217.x Rieder, C., Ritter, F., Raspe, M. and Peitgen, H.-O. (2008), Interactive Visualization of Multimodal Volume Data for Neurosurgical Tumor Treatment. Computer Graphics Forum, 27: 1055–1062. doi: 10.1111/j.1467-8659.2008.01242.x I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard. The manuscript in question describes different methods to visualize data acquired through MRI/fMRI scans in a three-dimensional manner. This is something that is sometimes done in current neuroimaging research, but that is rarely done in a standardized manner, which makes this guide timely and relevant. In many cases, researchers choose to use 2D images instead, which can sometimes distort or omit information, as fMRI depictions are derived from an inherently 3-dimensional signal. The current manuscript separately describes ways to visualize clusters of activation (i.e. activation as it would be found when running an fMRI experiment) and anatomical regions of interest. It also provides hyperlinks to download relevant visualization software. The author goes into sufficient detail to include, for example, information on price and OS compatibility of different software packages. Also, the text provides details about how to create the images within a particular software package, or functions that increase user efficiency. Information like this, in addition to several informative illustrations in the manuscript, will make this text particularly useful for many people working in neuroimaging, and I am convinced that the publication of this manuscript will lead to a fruitful online discussion about the best ways to visualize and report 3D brain data. The title, abstract, and structuring of the manuscript are well-written and appropriate for its purpose as a brief guide. Overall, this concise and informative guide is useful, interesting, and well-written. I recommend its indexing after some very minor comments (listed below) have been addressed to increase the readability of the manuscript. Minor suggestions: While the term ‘3D’ could be considered to be a household word, I would still recommend to spell it out as ‘three-dimensional (3D)’ or ‘3-dimensional (3D)’ the first time the term is used in the text. Likewise, the term ‘glass brain’ is intuitive, but not always used in the same way by all researchers. A quick description of the concept at the first mention of the term in the text would make the manuscript more accessible to the general reader. I have read this submission. I believe that I have an appropriate level of expertise to confirm that it is of an acceptable scientific standard.
  13 in total

1.  A 3-dimensional digital atlas of the ascending sensory and the descending motor systems in the pigeon brain.

Authors:  Onur Güntürkün; Marleen Verhoye; Geert De Groof; Annemie Van der Linden
Journal:  Brain Struct Funct       Date:  2012-02-25       Impact factor: 3.270

2.  Tract probability maps in stereotaxic spaces: analyses of white matter anatomy and tract-specific quantification.

Authors:  Kegang Hua; Jiangyang Zhang; Setsu Wakana; Hangyi Jiang; Xin Li; Daniel S Reich; Peter A Calabresi; James J Pekar; Peter C M van Zijl; Susumu Mori
Journal:  Neuroimage       Date:  2007-08-15       Impact factor: 6.556

3.  User-guided 3D active contour segmentation of anatomical structures: significantly improved efficiency and reliability.

Authors:  Paul A Yushkevich; Joseph Piven; Heather Cody Hazlett; Rachel Gimpel Smith; Sean Ho; James C Gee; Guido Gerig
Journal:  Neuroimage       Date:  2006-03-20       Impact factor: 6.556

4.  Imaging dopamine receptors in humans with [11C]-(+)-PHNO: dissection of D3 signal and anatomy.

Authors:  Andri C Tziortzi; Graham E Searle; Sofia Tzimopoulou; Cristian Salinas; John D Beaver; Mark Jenkinson; Marc Laruelle; Eugenii A Rabiner; Roger N Gunn
Journal:  Neuroimage       Date:  2010-06-30       Impact factor: 6.556

5.  Waxholm space: an image-based reference for coordinating mouse brain research.

Authors:  G Allan Johnson; Alexandra Badea; Jeffrey Brandenburg; Gary Cofer; Boma Fubara; Song Liu; Jonathan Nissanov
Journal:  Neuroimage       Date:  2010-07-01       Impact factor: 6.556

6.  A quantitative magnetic resonance histology atlas of postnatal rat brain development with regional estimates of growth and variability.

Authors:  Evan Calabrese; Alexandra Badea; Charles Watson; G Allan Johnson
Journal:  Neuroimage       Date:  2013-01-24       Impact factor: 6.556

7.  Common atlas format and 3D brain atlas reconstructor: infrastructure for constructing 3D brain atlases.

Authors:  Piotr Majka; Ewa Kublik; Grzegorz Furga; Daniel Krzysztof Wójcik
Journal:  Neuroinformatics       Date:  2012-04

8.  The INIA19 Template and NeuroMaps Atlas for Primate Brain Image Parcellation and Spatial Normalization.

Authors:  Torsten Rohlfing; Christopher D Kroenke; Edith V Sullivan; Mark F Dubach; Douglas M Bowden; Kathleen A Grant; Adolf Pfefferbaum
Journal:  Front Neuroinform       Date:  2012-12-06       Impact factor: 4.081

9.  A test-retest fMRI dataset for motor, language and spatial attention functions.

Authors:  Krzysztof J Gorgolewski; Amos Storkey; Mark E Bastin; Ian R Whittle; Joanna M Wardlaw; Cyril R Pernet
Journal:  Gigascience       Date:  2013-04-29       Impact factor: 6.524

10.  Large-scale automated synthesis of human functional neuroimaging data.

Authors:  Tal Yarkoni; Russell A Poldrack; Thomas E Nichols; David C Van Essen; Tor D Wager
Journal:  Nat Methods       Date:  2011-06-26       Impact factor: 28.547

View more
  16 in total

1.  Cortical complexity as a measure of age-related brain atrophy.

Authors:  Christopher R Madan; Elizabeth A Kensinger
Journal:  Neuroimage       Date:  2016-04-19       Impact factor: 6.556

2.  Predicting age from cortical structure across the lifespan.

Authors:  Christopher R Madan; Elizabeth A Kensinger
Journal:  Eur J Neurosci       Date:  2018-02-12       Impact factor: 3.386

3.  Age-related differences in the structural complexity of subcortical and ventricular structures.

Authors:  Christopher R Madan; Elizabeth A Kensinger
Journal:  Neurobiol Aging       Date:  2016-10-27       Impact factor: 4.673

4.  Network-level connectivity is a critical feature distinguishing dystonic tremor and essential tremor.

Authors:  Jesse C DeSimone; Derek B Archer; David E Vaillancourt; Aparna Wagle Shukla
Journal:  Brain       Date:  2019-06-01       Impact factor: 13.501

Review 5.  The ConTraSt database for analysing and comparing empirical studies of consciousness theories.

Authors:  Itay Yaron; Lucia Melloni; Michael Pitts; Liad Mudrik
Journal:  Nat Hum Behav       Date:  2022-02-21

6.  Functional Magnetic Resonance Imaging Studies in Sexual Medicine: A Primer.

Authors:  Colleen Mills-Finnerty; Eleni Frangos; Kachina Allen; Barry Komisaruk; Nan Wise
Journal:  J Sex Med       Date:  2022-04-11       Impact factor: 3.937

7.  Cognitive Aging and the Hippocampus in Older Adults.

Authors:  Andrew O'Shea; Ronald A Cohen; Eric C Porges; Nicole R Nissim; Adam J Woods
Journal:  Front Aging Neurosci       Date:  2016-12-08       Impact factor: 5.750

8.  Opportunities and considerations for visualising neuroimaging data on very large displays.

Authors:  Matthew B Wall; David Birch; May Y Yong
Journal:  F1000Res       Date:  2016-09-02

9.  Abnormal brain white matter microstructure is associated with both pre-hypertension and hypertension.

Authors:  Hideaki Suzuki; He Gao; Wenjia Bai; Evangelos Evangelou; Ben Glocker; Declan P O'Regan; Paul Elliott; Paul M Matthews
Journal:  PLoS One       Date:  2017-11-16       Impact factor: 3.240

10.  Age-related decrements in cortical gyrification: Evidence from an accelerated longitudinal dataset.

Authors:  Christopher R Madan
Journal:  Eur J Neurosci       Date:  2020-11-23       Impact factor: 3.386

View more

北京卡尤迪生物科技股份有限公司 © 2022-2023.