Batch effect

From 太極
Revision as of 16:30, 12 February 2024 by Brb (talk | contribs) (→‎ComBat)
Jump to navigation Jump to search

Merging two gene expression studies

Possible batch

STAT115 Chapter 6.5 Batch Effect Removal

Visualization

Chapter 2 Batch effect detection from the ebook Managing Batch Effects in Microbiome Data

ComBat

  • Adjusting batch effects in microarray expression data using empirical Bayes methods Johnson 2006. Note the term "ComBat" has not been used.
  • Statistics for Genomic Data Science (Coursera) and https://github.com/jtleek/genstats
  • Some possible batch variables: operators, runs, machines, library kits, laboratories.
  • sva::ComBat() function in sva package from Bioconductor. [math]\displaystyle{ \begin{align} Y_{ijg} = \alpha_g + X \beta_g + \gamma_{ig} + \delta_{ig} \epsilon_{ijg} \end{align} }[/math] where [math]\displaystyle{ X=X_{ij} }[/math] consists of covariates (eg biological) of scientific interests (e.g. Pathway activation levels in Zhang's 2018 simulation example), while [math]\displaystyle{ \gamma_{ig} }[/math] and [math]\displaystyle{ \delta_{ig} }[/math] characterize the additive and multiplicative batch effects of batch i for gene g. The error terms, [math]\displaystyle{ \epsilon_{ijg} }[/math], are assumed to follow a normal distribution with expected value of zero and variance [math]\displaystyle{ \sigma^2_𝑔 }[/math]. The batch corrected data is [math]\displaystyle{ \begin{align} \frac{Y_{ijg} - \hat{\alpha_g} - X \hat{\beta_g} - \hat{\gamma_{ig}}}{\hat{\delta_{ig}}} + \hat{\alpha_g} + X \hat{\beta_g}. \end{align} }[/math]
  • Alternative empirical Bayes models for adjusting for batch effects in genomic studies Zhang et al. BMC Bioinformatics 2018. The R package is sva and BatchQC from Bioconductor.
    • Reference batch adjustment: [math]\displaystyle{ \begin{align} Y_{ijg} = \alpha_{rg} + X \beta_{rg} + \gamma_{rig} + \delta_{rig} \epsilon_{ijg} \end{align} }[/math] where [math]\displaystyle{ \alpha_{rg} }[/math] is the average gene expression in the chosen reference batch (r). Furthermore, [math]\displaystyle{ \gamma_{rig} }[/math] and [math]\displaystyle{ \delta_{rig} }[/math] represent the additive and multiplicative batch differences between the reference batch and batch i for gene g. The error terms, [math]\displaystyle{ \epsilon_{ijg} }[/math], are assumed to follow a normal distribution with expected value of zero and a reference batch variance [math]\displaystyle{ \sigma^2_{𝑟𝑔} }[/math].
    • Mean-only adjustment for batch effects: [math]\displaystyle{ \begin{align} Y_{ijg} = \alpha_{g} + X \beta_{g} + \gamma_{ig} + \epsilon_{ijg} \end{align} }[/math]
  • Q: does data from each batch after correction have the same mean and variance?
    A: No, the data from each batch do not necessarily have the same mean and variance after correction by ComBat. The goal of ComBat is to adjust the data so that the batch effects are removed, making the data from different batches more comparable. However, this does not mean that the mean and variance of the data from each batch will be identical after correction.
    ComBat uses empirical Bayes methods (pooling information across genes to generate more robust estimates for gene-wise mean and variance batch effect) to estimate the parameters of the Gaussian distribution (mean and variance) for each batch. These parameters are then shrunk towards a common value. The data is adjusted based on these shrunken parameters. This process helps to remove the batch effects, but it does not force the data from each batch to have the same mean and variance.
    It’s important to note that while ComBat can effectively adjust for known batch effects, it may not completely eliminate all batch effects, especially if there are unknown or unmodeled batch effects present. Therefore, it’s always important to visually inspect the data and perform additional statistical tests to confirm the effectiveness of the batch correction.
  • Q: how to verify the batch effects have been corrected?
    • Visual Inspection: After batch correction, you can visually inspect the data using plots such as PCA plots or t-SNE plots. The samples from different batches should overlap with each other if the batch effects have been successfully removed.
    • Statistical Tests: You can perform statistical tests to check if the distribution of the data from different batches has become more similar after batch correction. For example, you can use tests such as the Kolmogorov-Smirnov test or the Anderson-Darling test.
    • Correlation and Linear Fit Plots: You can follow up the conditional plot with a linear correlation plot, a linear intercept plot, a linear slope plot, and a linear residual standard deviation plot. These four plots show the correlation, the intercept and slope from a linear fit, and the residual standard deviation for linear fits applied to each batch.
    • Benchmarking Metrics: Performance can be evaluated using benchmarking metrics including kBET, LISI, ASW, and ARI.
    • Differential Expression Analysis: Investigate the use of batch-corrected data to study differential gene expression.
  • Q: What is Bayesian estimation and Shrinkage. Empirical Bayes methods shrink the estimated parameters (eg sample mean from one gene for mean parameter) towards a common value (depending on the prior). Multiple genes/batches were used to estimate the hyperparameter. Th estimated hyperparameter was used to compute the posterior distribution of the unknown main parameter. See revisit and Understanding empirical Bayes estimation (using baseball statistics).
  • Naive method:
    # Step 1
    data <- matrix(rnorm(1000), nrow=100)
    batch <- factor(rep(c(1,2), each=50))
    # Step 2: estimate batch means
    batch_means <- tapply(1:nrow(data), batch, function(rows) colMeans(data[rows,]))
    # Step 3: estimate overall means
    overall_means <- colMeans(data)
    # Step 4: shrink batch means. Shrink the batch means towards the overall means. 
    # The degree of shrinkage could be determined empirically, or set to a fixed value.
    shrunken_means <- (batch_means + overall_means) / 2
    # Step 5; adjust data
    corrected_data <- data
    for(i in 1:ncol(data)) {
      corrected_data[,i] <- data[,i] - (shrunken_means[batch[i]] - overall_means[i])
    }
    
    
  • svg vignette example to remove the batch effect
    BiocManager::install("sva")
    library(sva)
    library(bladderbatch)
    data(bladderdata)
    pheno = pData(bladderEset)
    edata = exprs(bladderEset)
    batch = pheno$batch
    table(pheno$batch, pheno$cancer)
    table(batch)
    # batch
    #  1  2  3  4  5 
    # 11 18  4  5 19 
    
    modcombat = model.matrix(~1, data=pheno)
    combat_edata = ComBat(dat=edata, batch=batch, mod=modcombat, 
                          prior.plots=FALSE)
    # This returns an expression matrix, with the same dimensions 
    # as your original dataset (genes x samples).
    # mod: Model matrix for outcome of interest and other covariates besides batch
    # By default, it performs parametric empirical Bayesian adjustments. 
    # If you would like to use nonparametric empirical Bayesian adjustments, 
    # use the par.prior=FALSE option (this will take longer). 
    
    dim(edata)
    # [1] 22283    57
    combat_edata = ComBat(dat=edata, batch=batch, ref.batch=1)
    dim(combat_edata)
    # [1] 22283    57

    Does the result change if we use a subset of samples in combat but fix the ref.batch?
    Ans: the change is almost negligible

    (batch.sub <- batch %in% c(1,2,3,4))
    #  [1]  TRUE  TRUE  TRUE  TRUE  TRUE  TRUE  TRUE  TRUE FALSE  TRUE
    # [11] FALSE  TRUE FALSE  TRUE FALSE FALSE  TRUE  TRUE FALSE  TRUE
    # [21]  TRUE  TRUE  TRUE FALSE  TRUE  TRUE FALSE  TRUE  TRUE  TRUE
    # [31]  TRUE  TRUE  TRUE FALSE  TRUE  TRUE  TRUE  TRUE  TRUE  TRUE
    # [41] FALSE FALSE FALSE  TRUE FALSE  TRUE FALSE FALSE  TRUE  TRUE
    # [51] FALSE FALSE FALSE FALSE  TRUE  TRUE  TRUE
    combat_edata2 = ComBat(dat=edata[, batch.sub], batch=batch[batch.sub], ref.batch=1)
    range(combat_edata[, batch.sub] - combat_edata2)
    # [1] -0.00000000000004796163  0.00000000000003019807

    Does the result change if we use a subset of genes?
    Ans: yes, more appreciable depending on how much is the subset of genes.

    combat_edata3 = ComBat(dat=edata[1:20000,], batch=batch, ref.batch=1)
    range(combat_edata[1:20000,] - combat_edata3)
    # [1] -0.03803818  0.01561327
    combat_edata4 = ComBat(dat=edata[1:22000,], batch=batch, ref.batch=1)
    range(combat_edata[1:22000,] - combat_edata4)
    # [1] -0.006031209  0.012482280
  • ref.batch for reference-based batch adjustment. mean.only option if there is no need to adjust the variancec. Check out paper Alternative empirical Bayes models for adjusting for batch effects in genomic studies Zhang 2018. Figure 4 shows reference-based ComBat can clearly show the pathway activated samples in Batch 1 samples and show the true data pattern in Batch 2 samples from the simulated study (vs the original ComBat approach failed for both cases). In Figure 5 when we cluster genes using K-means, referenced-based Combat can better identify the role of DE or control genes (compared to the original ComBat method). In addition to the github reposition for the simulation R code, BatchQC::rnaseq_sim() can also do that.
  • Merging two gene-expression studies via cross-platform normalization by Shabalin et al, Bioinformatics 2008. This method (called Cross-Platform Normalization/XPN)was used by Ternès Biometrical Journal 2017.
  • Batch effect removal methods for microarray gene expression data integration: a survey by Lazar et al, Bioinformatics 2012. The R package is inSilicoMerging which has been removed from Bioconductor 3.4.
  • Question: Combine hgu133a&b and hgu133plus2. Adjusting batch effects in microarray expression data using empirical Bayes methods
  • Figure S1 shows the principal component analysis (PCA) before and after batch effect correction for training and validation datasets from another paper
  • Batch effects and GC content of NGS by Michael Love
  • 困扰的batch effect
  • Some note by Mikhail Dozmorov

CovBat

ComBat-Seq

svaseq

Applications

DESeq2

limma::removeBatchEffect()

ComBat or removebatcheffects (limma package)

ComBat or blocking in limma

batch effect : comBat or blocking in limma ?. The main difference between what Limma does and ComBat is that ComBat adjusts for differences in both the mean and variance differences across the batches, whereas Limma (I believe--Gordon please confirm) assumes that the batch variances are the same and only accounts for mean differences across the batches. So if there are large differences in batch variances, it might still be better to use ComBat. If there are not large variance differences, then Limma should be the best.

Repeat measure

Batch correction and repeated measures. You should just do a standard limma analysis as recommended. You should not use ComBat to pre batch-correct.

Comparisons of svaseq and Combat

Evaluation of Methods in Removing Batch Effects on RNA-seq Data. The results show the SVA method has the best performance, while the ComBat method over-corrects the batch effect.

HarmonizeR, proteomic

HarmonizR enables data harmonization across independent proteomic datasets with appropriate handling of missing values

MultiBaC- Multiomic Batch effect Correction

MultiBaC

BatchQC tool

BatchqcSummary.png BatchqcVariation.png

BatchqcDE.png BatchqcPCA.png

AMDBNorm - adjustment mean distribution-based normalization

AMDBNorm: an approach based on distribution adjustment to eliminate batch effects of gene expression data Zhang, Briefings in Bioinformatics 2021

AMDBNorm assumes that each sample in the biologic condition has the same distribution.

It selects a reference batch in advance, then

  1. aligns other batches with the distribution of the reference batch, and finally
  2. corrects the gene expression according to their mean values in the reference batch so that the expression of each gene in each batch is at the same level.

Evaluation methods

  • Visualization by principal component analysis
  • Evaluation of batch effects by BatchQC software
  • Principal variance component analysis
  • Hierarchical cluster analysis
  • Sensitivity analysis of AMDBNorm to reference batch
  • Adaptability analysis of batch correction methods to the new samples

TCGA

TCGAbatch_Correction()

Confounding factors

Impute

impute.knn

  • ?impute.knn impute.knn(data ,k = 10, rowmax = 0.5, colmax = 0.8, maxp = 1500, rng.seed=362436069)
    • rowmax: The maximum percent missing data allowed in any row (default 50%). For any rows with more than rowmax% missing are imputed using the overall mean per sample.
    • colmax: The maximum percent missing data allowed in any column (default 80%). If any column has more than colmax% missing data, the program halts and reports an error.
    Why are missing values on row 3 imputed by 0?
    BiocManager::install("impute")
    library(impute)
    
    set.seed(1)
    x <- matrix(rnorm(10*4), nr=10)
    x[1,1] <- NA
    x[2,1:2] <- NA
    x[3,1:3] <- NA
    x[1:9, 1] <- NA
    x2 <- impute.knn(x, k=2)
    # Error in impute.knn(x, k = 2) : 
    #   a column has more than 80 % missing values!
    
    set.seed(1)
    x <- matrix(rnorm(10*4), nr=10)
    x[1,1] <- NA
    x[2,1:2] <- NA
    x[3,1:3] <- NA
    x2 <- impute.knn(x, k=2)
    # Warning message:
    #   In knnimp(x, k, maxmiss = rowmax, maxp = maxp) :
    #   1 rows with more than 50 % entries missing;
    # mean imputation used for these rows
    names(x2)
    [1] "data"      "rng.seed"  "rng.state"
    head(x2$data)
    #            [,1]        [,2]        [,3]        [,4]
    # [1,]  0.1351965  1.51178117  0.91897737  1.35867955
    # [2,] -0.5629284  0.27448386  0.78213630 -0.10278773
    # [3,]  0.0000000  0.00000000  0.00000000  0.38767161
    # [4,]  1.5952808 -2.21469989 -1.98935170 -0.05380504
    # [5,]  0.3295078  1.12493092  0.61982575 -1.37705956
    
    set.seed(1)
    x <- 2^matrix(rnorm(10*4), nr=10)
    x[1,1] <- NA
    x[2,1:2] <- NA
    x[3,1:3] <- NA
    x2 <- impute.knn(x, k=2)
    head(x2$data)
    
    x3 <- impute.knn(x, k=2)
    identical(x2$data, x3$data) # true

    Another case that we are able to verify imputed for rows with more than rowmax% missing values. Two rows satisfy the condition. Note the missing value on row 3, column 3 is computed by excluding ALL rows satisfying the condition.

    set.seed(1)
    x <- matrix(rnorm(10*4), nr=10)
    x[1,1] <- NA
    x[2,1:2] <- NA
    x[3,1:3] <- NA
    
    x2 <- impute.knn(x, k=2, rowmax=.25)
    # Warning message:
    # In knnimp(x, k, maxmiss = rowmax, maxp = maxp) :
    #   2 rows with more than 25 % entries missing;
    #  mean imputation used for these rows
    head(x2$data)
    #            [,1]        [,2]        [,3]        [,4]
    # [1,]  0.1351965  1.51178117  0.91897737  1.35867955
    # [2,]  0.3714953  0.33998088  0.78213630 -0.10278773
    # [3,]  0.3714953  0.33998088 -0.27417921  0.38767161
    # [4,]  1.5952808 -2.21469989 -1.98935170 -0.05380504
    
    mean(x[,1], na.rm = T)
    # [1] 0.3714953  # same
    mean(x[,2], na.rm = T)
    # [1] 0.3399809  # same 
    mean(x[,3], na.rm = T)
    # [1] -0.1568108 # not the same
    
    mean(x[-c(2,3),3], na.rm = T)
    # [1] -0.2741792 # same
    
    x3 <- x2$data
    x3[1,1] <- NA
    dist(x3)  # find row1's 2 nearest neighbors are rows9 & 10.
    round(as.matrix(dist(x3))[1,], 2)
    #    1    2    3    4    5    6    7    8    9   10 
    # 0.00 2.17 2.23 5.70 3.21 2.95 2.96 3.27 1.82 1.39 
    mean(x[9:10,1])
    # [1] 0.1351965

Nearest-neighbor samples

kNN Imputation for Missing Values in Machine Learning. A new sample is imputed by finding the samples in the training set “closest” to it and averages these nearby points to fill in the value.

Proteomics data