This vignette describes the package Methylation-Aware Genotype Association with R (MAGAR) available from GitHub. MAGAR uses DNA methylation data obtained using the Illumina BeadArrays, and genotyping data from Illumina genotyping microarrays or whole genome sequencing to compute methylation quantitative trait loci (methQTL). The package provides mutliple flavors of linear modeling strategies to compute methQTL as statistically significant interactions between single nucleotide polymorphisms (SNPs) and changes in the DNA methylation state of individual CpGs. DNA methylation values at single CpGs are first summarized into correlation blocks, and a representative of this correlation block (tag-CpG) is used in the methQTL calling.
InstallationMAGAR can be installed using the basic Bioconductor installation functions.
if(!requireNamespace("BiocManager")){
install.packages("BiocManager")
}
if(!requireNamespace("MAGAR")){
BiocManager::install("MAGAR")
}
suppressPackageStartupMessages(library(MAGAR))
## No methods found in package 'oligoClasses' for request: 'mean' when loading 'crlmm'
Input data
MAGAR uses two types of data as input: DNA methylation data obtained using the Illumina Infinium BeadArrays or bisulfite sequencing and genotyping data obtained using genotyping microarrays or whole genome sequencing.
DNA methylation data (microarrays)MAGAR utilizes the widely used RnBeads software package for DNA methylation data import. Thus, MAGAR supports the various input options available in RnBeads, including a direct download from the Gene Expression Omnibus (GEO), IDAT, and BED files. For further options, we refer to the RnBeads vignette and documentation. In addition to the raw methylation data, a sample annotation sheet specifying the samples to be analyzed needs to be provided. The sheet contains a line for each sample and looks as follows:
SampleID,age,sex,barcode
Sample_1,14,f,209054857842_R01C01
Sample_2,42,f,209054857842_R02C01
Sample_3,45,m,209054857842_R03C01
For further details on the import process, we refer to the RnBeads vignette. Most importantly, analysis options need to be specified for the import and preprocessing modules of RnBeads. MAGAR provides a default setting, which is available in extdata/rnbeads_options.xml. You can use this file as a template for your own setting and then specify it to the package:
opts <- rnb.xml2options(system.file("extdata","rnbeads_options.xml",package="MAGAR"))
rnb.options(identifiers.column="geo_accession",
import.idat.platform="probes450")
xml.fi <- file.path(getwd(),"rnbeads_options.xml")
cat(rnb.options2xml(),file=xml.fi)
qtlSetOption(rnbeads.options = xml.fi)
Genotyping data PLINK files
MAGAR supports data that has already been processed by PLINK and that is available either in the form of binary .bed, .bim and .fam files, as .ped and .map, as variant calling files (.vcf), or as imputed files in the dosage format (.dos). For further processing, we use the command line tool PLINK, which is shipped together with MAGAR. However, this installation is only valid for Linux systems. For Windows and MacOS users, please install the PLINK tool from here and specify it using the option plink.path
. The sample identifier specified earlier also needs to match the sample IDs of the genotype calls. To import PLINK data, itâs best to store the .bim, .bed, and .fam files in a single folder (here plink_data
) and to specify the location of the folder to the package.
geno.dir <- "plink_data"
rnb.set <- load.rnb.set("rnb_set_dir")
s.anno <- "sample_annotation.csv"
data.loc <- list(idat.dir=rnb.set,geno.dir=plink.dir)
qtlSetOption(geno.data.type="plink",
meth.data.type="rnb.set")
meth.qtl <- doImport(data.location=data.loc,
s.anno=s.anno,
s.id.col="SampleID",
out.folder=getwd())
IDAT files
MAGAR also supports raw IDAT files and uses the CRLMM R-package, together with PLINK to perform genotype calling and data import. The package requires a single sample annotation sheet in the format described in the DNA methylation data section. In addition to the column names specified above, a column named GenoSentrixPosition has to be added, which specifies the IDAT file IDs.
SampleID,age,sex,barcode,GenoSentrixPosition
Sample_1,14,f,209054857842_R01C01,9701756058_R05C01
Sample_2,42,f,209054857842_R02C01,9701756058_R07C01
Sample_3,45,m,209054857842_R03C01,9742011016_R04C01
Imputation
Since Illumina SNP BeadArray data is typically imputed before further analysis, the package integrates a imputation functionality through the Michigan Imputation Server. Using the option setting described below, the package will automatically submit imputation jobs to the server and process the resulting files. In order to be able to perform computation on the server, an account is required. After the account is created, one has to request an API token in the user settings and specify it to MAGAR using the option imputation.user.token
. During the imputation process, the package will stall for a while and wait for the job to finish. After the job is completed, the package will prompt for entering the password send via e-mail to the user account. The imputation process has to be split according to chromosomes, which is why multiple e-mails will be send to the account, and the imputation process can take up to several days. However, after imputation, the imputed data will be available as PLINK files, such that the imputation has to be performed only once. For preprocessing the data for upload to the imputation server, the package requires the bgzip and tabix tools from the htslib package. Also see further options to configure the imputation jobs at the Michigan Imputation Server documentation:
qtlSetOption(
impute.geno.data=TRUE,
imputation.reference.panel="apps@hrc-r1.1",
imputation.phasing.method="shapeit",
imputation.population="eur"
)
Perform data import
The doImport
function requires the paths to the respective genotyping and DNA methylation data, as well as a sample annotation sheet as discussed earlier. In this vignette, we will describe the import of DNA methylation data in IDAT format and genotyping data as PLINK files. First, youâll have to specify the paths to the corresponding IDAT and plink files. Additionally, you have to specify the sample identifier column in the sample annotation sheet that determines the samples in both the genotyping and DNA methylation data. For larger files, we recommend to activate the option to store large matrices on disk rather than in main memory (hdf5dump
).
For imputed data, no further processing is performed on the genotyping data and the dosage values are used as they are:
idat.dir <- "idat_dir"
geno.dir <- "geno_dir"
anno.sheet <- "sample_annotation.tsv"
qtlSetOption(hdf5dump=TRUE)
imp.data <- doImport(data.location = c(idat.dir=idat.dir,geno.dir=geno.dir),
s.anno = anno.sheet,
s.id.col = "ID",
tab.sep = "\t",
out.folder = getwd())
Please note that the recode.allele.frequencies
option specifies, if, according to the cohort analyzed, SNP reference and alternative allele are to be recoded according to the allele frequencies in the available samples. Alternatively, a path to a local version of dbSNP (Sherry et al. 2001) can be provided through db.snp.ref
, and reference/alternative allele information will be automatically parsed from the database. This is especially crucial, if imputation is to be performed, since the Michigan Imputation Server is sensitive to reference mismatches.
Although MAGAR conceptually splits the methQTL calling into two steps ((i) compute correlation block, (ii) call methQTL per correlation block), only a single function call is needed. The function only requires the input methQTLInput
object produced in the previous step, but further options, such as covariates and the p-value cutoff can be directly specified as a function parameter, or as global parameters using ?qtlSetOption
.
imp.data <- loadMethQTLInput(system.file("extdata","reduced_methQTL",package="MAGAR"))
qtlSetOption(standard.deviation.gauss=100,
cluster.cor.threshold=0.75)
meth.qtl.res <- doMethQTL(imp.data,default.options=FALSE,p.val.cutoff=0.05)
## 2025-04-15 19:03:48 1.4 STATUS STARTED Imputation procedure knn
## 2025-04-15 19:03:48 1.4 STATUS COMPLETED Imputation procedure knn
##
## 2025-04-15 19:03:48 1.4 STATUS STARTED Computing methQTLs
## 2025-04-15 19:03:48 1.4 STATUS STARTED Computing methQTL for chromosome chr18
## 2025-04-15 19:03:48 1.4 STATUS STARTED Compute correlation blocks
## 2025-04-15 19:03:48 1.4 STATUS STARTED Compute correlation matrix
## 2025-04-15 19:03:48 1.5 STATUS COMPLETED Compute correlation matrix
## 2025-04-15 19:03:48 1.5 STATUS STARTED Compute pairwise distances
## 2025-04-15 19:03:50 1.5 STATUS COMPLETED Compute pairwise distances
## 2025-04-15 19:03:51 1.5 STATUS STARTED Weight distances
## 2025-04-15 19:03:51 1.5 STATUS COMPLETED Weight distances
## 2025-04-15 19:03:52 1.5 STATUS STARTED Compute graph
## 2025-04-15 19:03:52 1.5 STATUS COMPLETED Compute graph
## 2025-04-15 19:03:52 1.5 STATUS STARTED Compute clustering
## 2025-04-15 19:03:53 1.5 STATUS COMPLETED Compute clustering
## 2025-04-15 19:03:53 1.5 STATUS COMPLETED Compute correlation blocks
## Saving 7 x 5 in image
## 2025-04-15 19:03:53 1.5 STATUS STARTED Compute methQTL per correlation block
## 2025-04-15 19:03:53 1.5 STATUS STARTED Setting up Multicore
## 2025-04-15 19:03:53 1.5 INFO Using 1 cores
## 2025-04-15 19:03:53 1.5 STATUS COMPLETED Setting up Multicore
## 2025-04-15 19:04:11 1.8 STATUS COMPLETED Compute methQTL per correlation block
## 2025-04-15 19:04:11 1.8 STATUS COMPLETED Computing methQTL for chromosome chr18
## 2025-04-15 19:04:12 1.8 STATUS COMPLETED Computing methQTLs
We will now present the two steps of the methQTL calling procedure in more detail.
Compute CpG correlation blocksSince neighboring CpGs are often highly correlated, using each CpG independently as a potential methQTL candidate leads to many redundant results. We thus aimed to approximate DNA methylation haplotypes by determining highly correlated CpGs in close vicinity. The procedure itself is split into six steps, and is performed for each chromosome independently:
correlation.type
)cluster.cor.threshold
)standard.deviation.gauss
). Higher values for the standard deviation lead to a lower penalty on distal CpGs, thus the clusters will become larger.absolute.distance.cutoff
This will return a clustering according to the correlation structure between neighboring CpGs that we will later use for methQTL calling. Note that we used simultation experiments to determine the parameters for each data type individually. They will be automatically loaded for the dataset that is used and are:
cluster.cor.threshold
=0.2, standard.deviation.gauss
=5,000, absolute.distance.cutoff
=500,000cluster.cor.threshold
=0.2, standard.deviation.gauss
=3,000, absolute.distance.cutoff
=500,000cluster.cor.threshold
=0.2, standard.deviation.gauss
=250, absolute.distance.cutoff
=500,000From the list of correlation blocks, MAGAR computes methQTL interactions with all SNPs on the same chromosome. The process is split into three steps:
representative.cpg.computation
(default: row.medians).absolute.distance.cutoff
(default: 1,000,000) away from the representative CpGlinear.model.type
(default: classical.linear). Alternatively, fastQTL can be set as an option for meth.qtl.type
. This will tell the package to use the fastQTL software (Ongen et al. 2016).The meth.qtl.type
tells, how a methQTL interaction is defined and provides three options, in addition to the already mentioned fastQTL:
In the methQTL calling process, potential covariates can be specified using the option sel.covariates. We recommend to include at least age and sex as covariates, as they have a strong influence on the DNA methylation pattern.
Downstream analysis and interpretation How to use methQTLResultThe above procedure will create an object of class methQTLResult
, which contains the methQTL that are called in the previous step. To get a table of all the methQTL, you need to extract the information from the object. In the majority of the function calls below, there is the option type
, which takes on the values: * âSNPâ: To characterize the SNPs that influence any DNA methylation state * âCpGâ: To characterize the representative CpGs per correlation block that are influences by any SNP genotype * âcor.blockâ: To characterize all CpGs, which are part of a correlation block, whose representative CpG is influenced by any genotype
Furthermore, you can obtain genomic annotations for both the CpGs and the SNPs involved in the methQTL interactions:
result.table <- getResult(meth.qtl.res)
head(result.table)
anno.meth <- getAnno(meth.qtl.res,"meth")
head(anno.meth)
anno.geno <- getAnno(meth.qtl.res,"geno")
head(anno.geno)
For more detailed information about the output, also see the function getResultsGWASMap
.
To visualize methQTL, MAGAR provides some plotting functions. Most functions return an object of type ggplot
, which can be subsequently stored or viewed. Either all methQTL can be simultaneously visualized in a single plot, or a specific methQTL can be visualized:
result.table <- result.table[order(result.table$P.value,decreasing=FALSE),]
qtlPlotSNPCpGInteraction(imp.data,result.table$CpG[1],result.table$SNP[1])
## `geom_smooth()` using formula = 'y ~ x'
qtlDistanceScatterplot(meth.qtl.res)
Interpretation functions
The package provides a bunch of interpretation functions to characterize the detected methQTLs. This includes LOLA enrichment analysis(Sheffield and Bock 2016) (qtlLOLAEnrichment
), genomic annotation enrichment based on putative regulatory elements defined by the Ensembl Regulatory Build(Zerbino et al. 2015) (qtlAnnotationEnrichment
), enrichment analysis of different base substitutions in SNPs (qtlBaseSubstitutionEnrichment
), or TFBS motif enrichment using TFBSTools. Enrichment is compared for the methQTLs that are available in the provided methQTLResult
(for a single input), or to the overlapping QTLs for a list of methQTLResult
. The background of the enrichment is defined as all the SNPs/CpGs that have been used as input to the methQTL calling.
res <- qtlBaseSubstitutionEnrichment(meth.qtl.res)
Lists of methQTL results
Most of the functions discussed above either support a single methQTLResult
as input, or a list of such objects. In case a list is specified, the functions will typically overlap the methQTLs found and compare those with all SNPs/CpGs that have been used for methQTL calling. Additionally, there are functions that particularly work on a list of methQTLResult
objects and that perform overlapping, or determine the methQTLs specific to a dataset.
meth.qtl.res.2 <- loadMethQTLResult(system.file("extdata","MethQTLResult_chr18",package="MAGAR"))
meth.qtl.list <- list(First=meth.qtl.res,Second=meth.qtl.res.2)
qtlVennPlot(meth.qtl.list,out.folder=getwd())
## Loading required namespace: VennDiagram
qtlUpsetPlot(meth.qtl.list,type = "cor.block")
## 2025-04-15 19:04:14 1.9 STATUS STARTED Constructing universe
## 2025-04-15 19:04:14 1.9 STATUS STARTED Obtaining correlation blocks for object 1
## 2025-04-15 19:04:14 1.9 STATUS COMPLETED Obtaining correlation blocks for object 1
## 2025-04-15 19:04:14 1.9 STATUS STARTED Obtaining correlation blocks for object 2
## 2025-04-15 19:04:14 1.9 STATUS COMPLETED Obtaining correlation blocks for object 2
## 2025-04-15 19:04:14 1.9 STATUS STARTED Overlapping
## 2025-04-15 19:04:14 1.9 STATUS STARTED Obtaining correlation blocks for object 1
## 2025-04-15 19:04:14 1.9 STATUS COMPLETED Obtaining correlation blocks for object 1
## 2025-04-15 19:04:14 1.9 STATUS STARTED Obtaining correlation blocks for object 2
## 2025-04-15 19:04:14 1.9 STATUS COMPLETED Obtaining correlation blocks for object 2
## 2025-04-15 19:04:14 1.9 STATUS COMPLETED Overlapping
spec.first <- getSpecificQTL(meth.qtl.list$First,meth.qtl.list[-1])
Advanced configuration MAGAR options
MAGAR is a flexible software packages that allows for multiple flavors of methQTL analyses. Here, we present the different options that describe the analysis in a table and also discuss possible problems with choosing non-default options. Most options have reasonable default values that have been determined using a simulation experiment. For further information, see the documentation of qtlSetOption
vcftools.path
Path to an installation of VCFtools for handling VCF files for genotyping data plink.path
Path to an executable version of PLINK for processing genotyping data fast.qtl.path
Path to an executable version of FastQTL for calling methQTLs Only required, when meth.qtl.type='fastQTL'
bgzip.path
Path to an executable version of bgzip from the HTSlib package for compressing genomic data tabix.path
Path to an executable version of tabic from the HTSlib package for indexing genomic data External tool configuration rnbeads.options
Path to an XML file specifying the methylation data processing conducted through RnBeads rnbeads.report
Path to an existing directory, where the RnBeads report is to be stored rnbeads.qc
Flag indicating if the QC module of RnBeads is to be executed hardy.weinberg.p
Hardy-Weinberg test p-value cutoff as used for filtering SNPs in PLINK minor.allele.frequency
Minimum required minor allele frequency for a SNP to be considered in PLINK missing.values.samples
Maximum number of missing genotypes per SNP across the samples for a SNP to be considerd by PLINK plink.geno
Minimum genotyping rate for samples in PLINK n.permutations
Number of permutations used to correct for multiple testing in fastQTL Imputation options impute.geno.data
Flag indicating if genotyping data is to be executed or not imputation.user.token
User token generated by the Michigan Imputation server that allows for communicating with the Server API Needs to be generated at https://imputationserver.sph.umich.edu imputation.reference.panel
Reference panel used for the imputation See https://imputationserver.readthedocs.io/en/latest/api/ for further information imputation.phasing.method
Phasing method used for the imputation See https://imputationserver.readthedocs.io/en/latest/api/ for further information imputation.population
The population of the reference panel to be used for the imputation See https://imputationserver.readthedocs.io/en/latest/api/ for further information Import options meth.data.type
The type of methylation data that is about to be processed Accepted values are idat.dir
, data.dir
, data.files
, GS.report
, GEO
, or rnb.set
geno.data.type
The type of genotyping data that is about to be processed Accepted values are idat
or plink
Correlation Block Calling compute.cor.blocks
Flag indicating if correlation blocks are to be identified from the methylation data correlation.type
The type of correlation to be employed for computing correlation blocks Accepted values are pearson
, spearman
, and kendall
cluster.cor.threshold
The correlation threshold leading to an entry of zero in the correlation matrix standard.deviation.gauss
The standard deviation of the Gaussian that is used for weighting similarities according to the genomic distance absolute.distance.cutoff
The distance cutoff for the genomic distance. Higher distances lead to a zero in the similarity matrix max.cpgs
Maximum number of CpGs used for computing correlation blocks Depends on the main memory available MethQTL Calling meth.qtl.type
This options determines how a methQTL is defined Accepted values are oneVSall
, allVSall
, twoVSall
, or fastQTL
. linear.model.type
Determines how the linear model is defined for calling methQTLs Accepted values are categorical.anova
, classical.linear
, or fastQTL
representative.cpg.computation
Determines how a representative CpG is identified per correlation block, which is associated with the SNP genotype Accepted values are row.medians
, mean.center
, or best.all
n.prin.comp
Numeric value indicating how many of the PCs are to be used as covariates in the identification of methQTLs General options hdf5dump
Flag indicating if large matrices are to be stored on disk rather than in main memory using the HDF5Array
package Consider further options for HDF5Array
such as setHDF5DumpDir
and setHDF5DumpFile
db.snp.ref
A path to a downloaded version of dbSNP. With dbSNP, SNP identifiers can be extracted and annotated SNPs can be removed from the methylation data dbSNP should be in vcf.gz format and can be downloaded from https://ftp.ncbi.nih.gov/snp/organisms/human_9606_b150_GRCh37p13/VCF/ cluster.config
Configuration for a high perfomance computing cluster. Currently supported are SLURM and SGE. cluster.architecture
String indicating which high performance computing architecture is used Currently supported are âslurmâ and âsgeâ recode.allele.frequencies
Flag indicating if the allele frequencies are to be computed from the genotype data. Employ MAGAR on a scientific compute cluster
MAGAR can automatically distribute jobs across a high performance compute cluster, which has been setup using the Sun Grid Engine (SGE) technology. You can pass the option cluster.submit
to doMethQTL
and thus activate the cluster submission. Note that youâll also have to specify a path to an executable Rscript and potentially specify resource requirements using the option setting cluster.config
.
qtlSetOption(cluster.config = c(h_vmem="60G",mem_free="20G"))
qtlSetOption(rscript.path = "/usr/bin/Rscript")
meth.qtl.res <- doMethQTL(meth.qtl = imp.data,
cluster.submit = T)
References
Ongen, Halit, Alfonso Buil, Andrew Anand Brown, Emmanouil T. Dermitzakis, and Olivier Delaneau. 2016. âFast and efficient QTL mapper for thousands of molecular phenotypes.â Bioinformatics 32 (10): 1479â85. https://doi.org/10.1093/bioinformatics/btv722.
Sheffield, Nathan C., and Christoph Bock. 2016. âLOLA: enrichment analysis for genomic region sets and regulatory elements in R and Bioconductor.â Bioinformatics 32 (4): 587â89. https://doi.org/10.1093/bioinformatics/btv612.
Sherry, S. T., M. H. Ward, M. Kholodov, J. Baker, L. Phan, E. M. Smigielski, and K. Sirotkin. 2001. âdbSNP: the NCBI database of genetic variation.â Nucleic Acids Research 29 (1): 308â11. https://doi.org/10.1093/nar/29.1.308.
Zerbino, Daniel R., Steven P. Wilder, Nathan Johnson, Thomas Juettemann, and Paul R. Flicek. 2015. âThe Ensembl Regulatory Build.â Genome Biology 16 (1): 1â8. https://doi.org/10.1186/s13059-015-0621-5.
RetroSearch is an open source project built by @garambo | Open a GitHub Issue
Search and Browse the WWW like it's 1997 | Search results from DuckDuckGo
HTML:
3.2
| Encoding:
UTF-8
| Version:
0.7.4