| Literature DB >> 21743807 |
Daniel S Marcus1, John Harwell, Timothy Olsen, Michael Hodge, Matthew F Glasser, Fred Prior, Mark Jenkinson, Timothy Laumann, Sandra W Curtiss, David C Van Essen.
Abstract
The Human Connectome Project (HCP) is a major endeavor that will acquire and analyze connectivity data plus other neuroimaging, behavioral, and genetic data from 1,200 healthy adults. It will serve as a key resource for the neuroscience research community, enabling discoveries of how the brain is wired and how it functions in different individuals. To fulfill its potential, the HCP consortium is developing an informatics platform that will handle: (1) storage of primary and processed data, (2) systematic processing and analysis of the data, (3) open-access data-sharing, and (4) mining and exploration of the data. This informatics platform will include two primary components. ConnectomeDB will provide database services for storing and distributing the data, as well as data analysis pipelines. Connectome Workbench will provide visualization and exploration capabilities. The platform will be based on standard data formats and provide an open set of application programming interfaces (APIs) that will facilitate broad utilization of the data and integration of HCP services into a variety of external applications. Primary and processed data generated by the HCP will be openly shared with the scientific community, and the informatics platform will be available under an open source license. This paper describes the HCP informatics platform as currently envisioned and places it into the context of the overall HCP vision and agenda.Entities:
Keywords: Human Connectome Project; XNAT; brain parcellation; caret; connectomics; diffusion imaging; network analysis; resting state fMRI
Year: 2011 PMID: 21743807 PMCID: PMC3127103 DOI: 10.3389/fninf.2011.00004
Source DB: PubMed Journal: Front Neuroinform ISSN: 1662-5196 Impact factor: 4.081
Figure 1HCP subject workflow.
Figure 2The Connectome UI. (Left) This mockup of the Visualization & Discovery track illustrates key concepts that are being implemented, including a faceted search interface to construct subject groups and an embedded version of Connectome Workbench. Both the search interface and Workbench view are fed by ConnectomeDB's open API. (Right) This mockup of the Download track illustrates the track's emphasis on guiding users quickly to standard download packages and navigation to specific data.
Figure 3ConnectomeDB architecture, including data transfer components. ConnectomeDB will utilize the Tomcat servlet container as the application server and use the enterprise grade, open source PostgreSQL database for storage of non-imaging data, imaging session meta-data, and system data. Actual images and other binary content are stored on a file system rather than in the database, improving performance and making the data more easily consumable by external software packages.
The HCP computing infrastructure.
| Component | Device | Notes |
|---|---|---|
| Virtual cluster | 2 Dell PowerEdge R610s managed byVMWare ESXi | Additional nodes will be added in years 3 and 5. Dynamically expandable using NIAC cluster. |
| Web servers | VMs running Tomcat 6.0.29 and XNAT 1.5 | Load-balanced web servers host XNAT system and handle all API requests. Monitored by Pingdom and Google Analytics. |
| Database servers | VMs running Postgres 9.0.3. | Postgres 9 is run in synchronous multi-master replication mode, enabling high availability and load balancing. |
| Compute Cluster | VMs running Sun Grid Engine-based queuing. | Executes pipelines and on-the-fly computations that require short latencies. |
| Data storage | Scale-out NAS (Vendor TBD) | Planned 1 PB capacity will include tiered storage pools and 10Gb connectivity to cluster and HPCS. |
| Load balancing | Kemp Technologies LoadMaster 2600 | Distributes web traffic across multiple servers and provides hardware-accelerated SSL encryption |
| HPCS | IBM system in WU's CHPC | The HPC will execute computationally intensive processing including “standard” pipelines and user-submitted jobs. |
| DICOM gateway | Shuttle XS35-704 Intel Atom D510 | The gateway uses CTP to manage secure transmission of scans from UMinn scanner to ConnectomeDB. |
| Elastic computing and storage | Partner institutions, cloud computing | Mirror data sites will ease bottlenecks during peak traffic periods. Elastic computing strategies will automatically detect stress on compute cluster and recruit additional resources. |
The web servers, database servers, and compute cluster are jointly managed as a single VMware ESXi cluster for efficient resource utilization and high availability. The underlying servers each include 48-GB memory and dual 6-core processors. Each node in the VMware cluster is redundantly tied back in to the storage system for VM storage. All nodes run 64-bit CentOS 5.5. The HPCS includes an iDataPlex cluster (168 nodes with dual quad core Nehalem processors and 24-GB RAM), an e1350 cluster (7 SMP servers, each with 64 cores and 256-GB RAM), a 288-port Qlogic Infiniband switch to interconnect all processors and storage nodes, and 9 TB of high-speed storage. Connectivity to the system is provided by a 4 × 10 Gb research network backbone.
Figure 4HCP data distribution tiers.
Figure 5Connectome Workbench visualization of the inflated atlas surfaces for the left and right cerebral hemispheres plus the cerebellum. Probabilistic architectonic maps are shown of area 18 on the left hemisphere and area 2 on the right hemisphere.