Literature DB >> 30891564

A brief procedure for big data analysis of gene expression.

Kewei Wang1,2,3,4,5,6, Wenji Wang1,2,3,4, Mang Li1,2,3,4.   

Abstract

There are a lot of biological and experimental data from genomics, proteomics, drug screening, medicinal chemistry, etc. A large amount of data must be analyzed by special methods of statistics, bioinformatics, and computer science. Big data analysis is an effective way to build scientific hypothesis and explore internal mechanism. Here, gene expression is taken as an example to illustrate the basic procedure of the big data analysis.

Entities:  

Keywords:  PCA analysis; big data analysis; cluster analysis; microarray; regression model

Year:  2018        PMID: 30891564      PMCID: PMC6388068          DOI: 10.1002/ame2.12028

Source DB:  PubMed          Journal:  Animal Model Exp Med        ISSN: 2576-2095


INTRODUCTION

Gene expression data can be originated from different techniques such as quantitative real‐time PCR (qRT‐PCR), microarray, ChIP assay, ChIP‐on‐chip, high‐throughput ChIP‐sequencing, etc. Levels of the gene expression are calculated based on relative quantity of house‐keeping genes (ie, GAPDH, β‐actin, cyclophilin A1) or absolute magnitude subsequent to a standard curve. The gene expression of special proteins may be typical marker under physiological and pathological conditions. In the field of clinical medicine, biomarkers such as calcitonin for medullary thyroid carcinoma, alpha fetoprotein for hepatocellular carcinoma, glial fibrillary acidic protein for glioma, carcinoma antigen 15‐3 for breast cancer, and so on, are often utilized for diagnosis and prognostic evaluation of relevant tumors.1, 2, 3, 4 In addition, F13A1 gene is for screening abnormal bleeding risk.5 Parkinson's disease may have a high level of tissue transglutaminase.6 Gene T235 is a marker for the persistent microalbuminuria in children and adolescents with type 1 diabetes mellitus. So far, enough evidence has demonstrated differential levels of gene expression are closely associated with their functionality under pathophysiological status. However, how to form a scientific hypothesis based on existing data? How to choose suitable datasets and further to set up a relationship among different variables? It needs not only professional knowledge and judgment, but also mathematical theory and statistical methods. The present study introduces a brief procedure to perform big data analysis of gene expression.

DATA TYPES OF GENE EXPRESSION

The gene expression can be detected or measured via routine PCR or qRT‐PCR. One or more genes are determined at one time, based on designed primers. Relative quantification of the gene expression is usually normalized to the level of housekeeping gene with the ∆∆Ct method. Microarray utilizes single DNA chip to detect a lot of gene expression simultaneously. The principle of DNA microarray is to hybridize unknown DNA sequence with a set of probe oligodeoxynucleotides. The probe nucleotides are immobilized on hard surfaces such as glass in array way, and then were hybridized with fluorescent target genes. The gene expression is determined via fluorescence intensity/position of target nucleotides in the microarray. ChIP‐sequencing or ChIP‐seq is a technique to analyze the binding sites of DNA‐associated nuclear proteins, to investigate interactions between nuclear proteins and specific DNA sequence, and to identify functional role of transcription factors. ChIP‐Seq data are derived from the DNA sequencing through a series of operation. Now, the ChIP‐Seq has already replaced the previous ChIP‐on‐chip method for the interactive study between nuclear proteins and genomic DNA. Owing to differences in gene detection methods, there are multiple data sources of gene expression, for example, DNA microarray, Chip‐seq data (Figure 1). The first step for gene expression analysis is to cluster gene data with similar characteristics into different groups for further investigation.7 There are a few terms that explain data types of gene expression in Geo database as follows:
Figure 1

Data sources of gene expression derived from different detection methods

GSM (GEO sample) stands for the experimental data of a single sample. GDS (GEO dataset) is a collection of GSM that is arranged manually on a topic. Thus, GSM and GDS share the same platform. GSE (GEO series) includes a multiple experiment in a research project. It may use different platforms. GPL (GEO platform) is a chip platform, such as Affymetrix, Sentrix, Illumina, Aglent, etc. Data sources of gene expression derived from different detection methods

DATA PREPROCESSING

Levels of gene expression are much varied due to their diversity in nature/function. Raw data from different databases have to be reorganized by gene ID and symbol after having matched microarray annotation table (Table 1). If the raw data are used for direct comparison, it can overemphasize the role of the high abundance genes during the comprehensive analysis, and/or may weaken the function of the low gene‐expression levels. Therefore, in order to ensure the reliability of the result, the original data need to be normalized.8, 9 There are many kinds of data normalization methods, including min‐max normalization, log function conversion, arc tangent function conversion, z‐score normalization, and fuzzy quantization. They may be linear type (ie, M ± SD) and curve type (ie, seminormal distribution). Different normalization methods have diverse effects on the evaluation of resultant data.10, 11, 12 Unfortunately, there is no general rule to follow in the selection of data normalization methods. The min‐max normalization, also called standardization of deviations, is the linear transformation of the original data, and the result falls to the [0, 1] interval. The log function conversion can be completed via x* = log 10 (x). A problem is that the result does not necessarily fall on the [0, 1] interval (Table 2). It should be divided by log10 (max). The atan function transformation uses an inverse tangent function to achieve the normalization of data. The data less than 0 will be mapped to the [−1, 0] interval. The most common method is the z‐score normalization (or zero‐mean normalization) (Table 3), which is used as standard method in spss package (The IBM SPSS® Software, Armonk, NY, USA). The processed data conforms to the normal distribution, that is, the mean is 0 and the SD is 1. Its conversion function contains μ for the mean value and σ for the SD of all sample data.
Table 1

Different chip categories of GEO microarrays in NCBI database

DatasetsYearsGene numberNames
1200733297Aflymetrix Human Gene 1.0 ST array
2200354675Aflymetrix Human Genome U133 Plus 2.0 Array
3200222283Aflymetrix Human Genome U133A Array
420028793Affymetrix Human HG‐Focus Target Array
5200720228Agilent‐012097 Human 1A Microarray (V2) G4110B
6200645220Agilent‐014850 Whole Human Genome Microarray 4×44K G4112F
7200820589Sentrix Human Ref‐8 v2 Expression BeadChip
8200824526Illumina HumanRef‐8 v3.0 expression beadchip
Table 2

Logarithmic transformed data (log e ratio, reference series: GSE9539)

Gene4/100 Fold8/100 Fold12/100 Fold24/100 Fold4/200 Fold8/200 Fold12/200 Fold
BAD0.93300.95600.93900.99300.96000.97700.9800
BAX1.07100.96401.00700.93901.09700.94300.9870
BCL21.07300.94900.90601.06201.23600.78701.2920
BCL2A10.93201.36200.94100.92800.90900.91100.8130
BCL2L111.07400.97901.18100.96700.98801.01601.1500
BCL2L131.12900.97201.09901.01701.10800.99301.0610
BCL2L21.13001.04801.04101.06501.11201.05301.0290
BIK1.04400.95600.97701.01200.99200.95601.0480
BOK0.90000.93500.93000.95700.96200.93200.8810
Table 3

Transformed data by zero‐mean normalization

Gere symbol153‐T0‐1153‐T0‐2153‐T0‐3153‐T0.5‐1153‐T0.5‐2153‐T0.5‐3153‐T1.S‐1153‐T1.S‐1
TNFS F10−0.5237−0.6558−0.4203−0.5577−0.4949−0.61402.52253.2670
TNF0.2021−0.31060.0823−0.24350.2932−0.4927−0.52151.5486
TNFSF 120.35961.05380.19450.3850−0.03401.8283−0.97791.9087
FAS−1.0077−0.0496−0.54110.0521−0.8960−0.1018−1.3602−1.1021
TNFRSF 10B1.87971.68440.90531.95021.02381.23440.0413−0.6137
TNFRSF 10A−0.98250.0200−0.6054−0.5993−1.0315−0.63300.1672−0.5839
FADD0.19150.26940.8454−0.5752−0.4390−0.5752−0.8126−0.9527
TNFRSF 1A0.04430.72250.07730.07330.3874−0.4379−0.49590.0333
TNFRSF 10D0.49910.64491.10890.36000.68910.5102−0.2631−0.2520
TNFRSF 10C−0.5613−0.09151.80410.2629−0.5284−0.66030.66671.7877
TRAF2−0.3673−0.5960−0.44350.6127−0.37280.8686−0.4163−1.6087
TRADD−0.1146−0.9489−0.9365−1.2715−1.5972−0.85900.20481.0081
TNFRS F11B−0.8241−0.4364−0.5804−0.5804−1.1564−0.58040.17290.1729
TRAF10.04940.7638−0.0038−1.4858−1.2730−1.72900.0190−0.6498
Different chip categories of GEO microarrays in NCBI database Logarithmic transformed data (log e ratio, reference series: GSE9539) Transformed data by zero‐mean normalization

MULTIVARIATE DATA ANALYSIS

There are different statistical methods for big data analysis. Several techniques that are generally applied for microarray gene expression are summarized as follows.

Correlation analysis

It describes the correlation between two or more than two random variables, for example, the relationship between body height and weight, and relative humidity between the air and rainfall. When abovementioned variables are plotted in the Cartesian coordinate system, there is a set of points called “scatter plot.” The correlation between the two variables is expressed by correlation coefficient r. The correlation coefficient r is any value between −1 and 1.13, 14 The correlation between one variable x 0 and a set of variables (x 1, x 2, … x ) is measured by multicorrelation coefficient. The range of the multicorrelation coefficient is [0, 1]. In the case of multivariable, the net correlation between two variables is reflected by partial correlation coefficient when the effect of other variables is controlled or considered as a constant.15, 16 For instance, there are three variables X 1, X 2, and X 3. The partial correlation coefficient r13.2 represents the linear correlation between variable X 1 and X 3 after the influence of variable X 2 is thought of as a constant. The partial correlation coefficient can reflect the relationship between the two variables more accurately than the simple linear correlation coefficient.

Cluster analysis

The goal of cluster analysis is to collect data on similar basis for classification (Figure 2).17, 18 In different applications, many clustering techniques have been developed, which are used to describe data, to measure the similarity between different data sources, and to classify data sources into different clusters. The statistical methods for the cluster analysis include system clustering, decomposition, addition, dynamic clustering, ordered sample clustering, overlapping clustering, and fuzzy clustering. Traditional clustering algorithms can be classified into five categories: partitioning, hierarchical, density‐based, grid based, and model‐based. Clustering analysis tools, such as k‐mean and k‐center algorithm, have been added to many famous packages such as spss (The IBM SPSS® Software) and sas (Cary, NC, USA).
Figure 2

Cluster analysis of gene expression. Different colors represent clusters of similar characteristics based on geometric distance

Cluster analysis of gene expression. Different colors represent clusters of similar characteristics based on geometric distance

Principal components analysis

In many cases, there is a certain correlation between the two variables. It can be explained that the two variables share common data information or have a overlap of datasets. The principal components analysis (PCA) is a technology for analyzing and simplifying datasets (Figure 3).19, 20 There is more information about PCA theory and application in the platforms “Principal Component Analysis (PCA) in R” and “Principal Component Analysis in Python” (website https://datascienceplus.com/principal-component-analysis-pca-in-r/; https://plot.ly/ipython-notebooks/principal-component-analysis/).19, 20, 21 The PCA method is usually used to reduce the dimension of dataset, while maintaining the greatest contribution of the other side in the dataset. This is achieved by preserving low‐order principal components and ignoring higher order principal components. Such low‐order components often retain the most important aspect of data. The PCA technique can take out as many of the less comprehensive variables as possible to reflect the information of the original variables. However, this is not necessary and depends on the specific application. Moreover, since PCA relies on the given data, the accuracy of the data has a great impact on the analysis results. The most classical approach in PCA analysis is to express the variance of F1. The F1 selected in all linear combinations should be the largest variance, so the F1 is the first principal component. If the first principal component is not enough to represent the information of the original variables, then consider selecting F2 for the second linear combinations, expressed in the mathematical Cov (F1, F2) = 0.
Figure 3

PCA analysis of gene expression. The horizontal and vertical coordinates represent two‐dimensional distribution of the first two components. The number in the figure stands for different samples

PCA analysis of gene expression. The horizontal and vertical coordinates represent two‐dimensional distribution of the first two components. The number in the figure stands for different samples

Regression model

Representative gene variables are chosen out of above‐mentioned original data following a series of statistical procedure. They are utilized to establish mathematical model and to characterize essential features of gene expression (Figure 4). Mathematical modeling is the process of describing actual phenomena in mathematical language.21, 22 The actual phenomena here contain specific natural phenomenon such as free falling, and abstractive phenomenon such as the value tendency of a customer to a certain commodity. The biological mathematical (BioMath) model is a strict language to make the description of all kinds of phenomena into a much scientific, logical, objective, and repeatable structure.23 The description here includes not only the external form and internal mechanism, but also the prediction, experiment, and explanation of the actual phenomena. Sometimes, we need to do some experiments, but these experiments often use abstract models as a substitute for the actual objects, and the experiment itself is a theoretical substitute for the physical operation. Therefore, the BioMath model is a simplification of real things. The establishment of mathematical models is the process of simplifying and abstracting complex and practical problems into a rational mathematical structure. Mathematical modeling is the bridge linking gene expression and biological/medical problems. Based on the hypothesis, we should collect original data of the gene expression, observe and investigate the inherent characteristics and inherent laws, ascertain the quantitative relationship among variables, to calculate and estimate all the parameters of the model, and then to verify the accuracy, rationality, and applicability of the model.
Figure 4

Sigmoid curve is a typical pattern of biological response as reflected by possibility range [0, 1]. Logistic regression or logit regression model is able to estimate the probability of a binary response based on one or more independent variables.

Sigmoid curve is a typical pattern of biological response as reflected by possibility range [0, 1]. Logistic regression or logit regression model is able to estimate the probability of a binary response based on one or more independent variables. In summary, big data accumulated through diverse methods can show different data types. Representative variables are screened out of multiple categories via professional knowledge and statistical methods. The representative variables may be fitted into a suitable mathematical model for the evaluation of pathophysiological process, which can be utilized in clinical diagnosis and treatment.

CONFLICT OF INTEREST

None.
  2 in total

1.  ATP6V1G3 Acts as a Key Gene in Recurrent Spontaneous Abortion: An Integrated Bioinformatics Analysis.

Authors:  Yihong Chen; Jifen Hu
Journal:  Med Sci Monit       Date:  2020-10-08

2.  The molecular regulation of autophagy in antimicrobial immunity.

Authors:  Chuan Qin; Yalan Lu; Lin Bai; Kewei Wang
Journal:  J Mol Cell Biol       Date:  2022-07-29       Impact factor: 8.185

  2 in total

北京卡尤迪生物科技股份有限公司 © 2022-2023.