synapter 2.20.1
synapter is free and open-source software. If you use it, please support the project by citing it in publications:
Nicholas James Bond, Pavel Vyacheslavovich Shliaha, Kathryn S. Lilley, and Laurent Gatto. Improving qualitative and quantitative performance for MS\(^E\)-based label free proteomics. J. Proteome Res., 2013, 12 (6), pp 2340–2353
For bugs, typos, suggestions or other questions, please file an issue
in our tracking system (https://github.com/lgatto/synapter/issues)
providing as much information as possible, a reproducible example and
the output of sessionInfo()
.
If you don’t have a GitHub account or wish to reach a broader audience for general questions about proteomics analysis using R, you may want to use the Bioconductor support site: https://support.bioconductor.org/.
Here we describe the new functionality implemented in synapter 2.0. Namely this vignette covers the utilisation of the new 3D grid search, the fragment matching, intensity modeling and correction of detector saturation.
The synapter2 workflow is similar to the old one in synapter1.
First it is necessary to use PLGS to create the csv (and xml) files.
Therefore we refer the reader to the default synapter vignette,
available online
and with vignette("synapter", package = "synapter")
.
In contrast to the original workflow the final_fragment.csv
file for the
identification run and a Spectrum.xml
file for the quantification run are
needed if the fragment matching should be applied.
Subsequently the original workflow is enhanced by the new
3D grid search and the intensity modeling.
Afterwards the fragment matching could be applied.
MSnbase (Gatto and Lilley 2012) is used for further analysis.
The new synapter adds
synapter/PLGS consensus filtering
and the detector saturation correction for MSnSet
s.
Synapter
objectTo demonstrate a typical step-by-step workflow we use example data
that are available on http://proteome.sysbiol.cam.ac.uk/lgatto/synapter/data/.
There is also an synobj2
object in synapterdata which
contains the same data.
The Synapter
constructor uses a named list
of input files. Please note that
we add identfragments
(final_fragment.csv
) and
quantspectra
(Spectrum.xml
) because we want to apply the fragment matching
later.
## Please find the raw data at:
## http://proteome.sysbiol.cam.ac.uk/lgatto/synapter/data/
library("synapter")
inlist <- list(
identpeptide = "fermentor_03_sample_01_HDMSE_01_IA_final_peptide.csv.gz",
identfragments = "fermentor_03_sample_01_HDMSE_01_IA_final_fragment.csv.gz",
quantpeptide = "fermentor_02_sample_01_HDMSE_01_IA_final_peptide.csv.gz",
quantpep3d = "fermentor_02_sample_01_HDMSE_01_Pep3DAMRT.csv.gz",
quantspectra = "fermentor_02_sample_01_HDMSE_01_Pep3D_Spectrum.xml.gz",
fasta = "S.cerevisiae_Uniprot_reference_canonical_18_03_14.fasta")
synobj2 <- Synapter(inlist, master=FALSE)
synobj2
## Object of class "Synapter"
## Class version 2.0.0
## Package version 1.99.0
## Data files:
## + Identification pep file: fermentor_03_sample_01_HDMSE_01_IA_final_peptide.csv.gz
## + Identification Fragment file: fermentor_02_sample_01_HDMSE_01_Pep3D_Spectrum.xml.gz
## + Quantitation pep file: fermentor_02_sample_01_HDMSE_01_IA_final_peptide.csv.gz
## + Quantitation Pep3DAMRT file: fermentor_02_sample_01_HDMSE_01_Pep3DAMRT.csv.gz
## + Quantitation Spectrum file: fermentor_02_sample_01_HDMSE_01_Pep3D_Spectrum.xml.gz
## + Fasta file: S.cerevisiae_Uniprot_reference_canonical_18_03_14.fasta
## Log:
## [1] "Instance created on Tue Apr 4 09:01:34 2017"
## [2] "Read identification peptide data [11471,16]"
## [ 10 lines ]
## [13] "Read identification fragment data [7078]"
## [14] "Read quantitation spectra [36539]"
The first steps in each synapter analysis are filtering by peptide sequence, peptide length, ppm error and false positive rate.
Here we use the default values for each method. But the accompanying plotting methods should be used to find the best threshold:
filterUniqueDbPeptides(synobj2,
missedCleavages=0,
IisL=TRUE)
filterPeptideLength(synobj2, l=7)
plotFdr(synobj2)
filterQuantPepScore(synobj2, method="BH",
fdr=0.05)
filterIdentPepScore(synobj2, method="BH",
fdr=0.05)
par(mfcol=c(1, 2))
plotPpmError(synobj2, what="Ident")
plotPpmError(synobj2, what="Quant")
par(mfcol=c(1, 1))
filterQuantPpmError(synobj2, ppm=20)
filterIdentPpmError(synobj2, ppm=20)
plotPepScores(synobj2)
filterIdentProtFpr(synobj2, fpr=0.05)
filterQuantProtFpr(synobj2, fpr=0.05)
Next we merge the identified peptides from the identification run and quantification run and build a LOWESS based retention time model to remove systematic shifts in the retention times. Here we use the default values but as stated above the plotting methods should be used to find sensible thresholds.
mergePeptides(synobj2)
plotRt(synobj2, what="data")
setLowessSpan(synobj2, span=0.05)
modelRt(synobj2)
par(mfcol=c(1, 2))
plotRtDiffs(synobj2)
plotRt(synobj2, what="model", nsd=1)
par(mfcol=c(1, 1))
plotFeatures(synobj2, what="all", ionmobility=TRUE)
To find EMRTS (exact m/z-retention time pairs) we try are running a grid search to find the best retention time tolerance and m/z tolerance that results in the most correct one-to-one matching in the merged (already identified) data. If the identification and quantitation run are HDMS\(^E\) data we could use the new 3D grid search that looks for the best matching in the retention time, m/z and ion mobility (drift time) domain to increase the accuracy. If one or both datasets are MS\(^E\) data it falls back to the traditional 2D grid search.
searchGrid(synobj2,
imdiffs=seq(from=0.6, to=1.6, by=0.2),
ppms=seq(from=2, to=20, by=2),
nsds=seq(from=0.5, to=5, by=0.5))
setBestGridParams(synobj2)
findEMRTs(synobj2)
plotEMRTtable(synobj2)
For the details of the fragment matching procedure we refer to the
fragment matching vignette that is available
online
and with vignette("fragmentmatching", package = "synapter")
.
Briefly we compare the fragments of the identification run with the spectra from
the quantification run and remove entries where there are very few/none common
peaks/fragments between them.
First we starting by removing less intense fragments and peaks.
filterFragments(synobj2,
what="fragments.ident",
minIntensity=70)
filterFragments(synobj2,
what="spectra.quant",
minIntensity=70)
Next we look for common peaks via fragmentMatching
:
fragmentMatching(synobj2)
## Warning in object$fragmentMatching(verbose = verbose): FragmentMatching ppm
## tolerance undefined. Setting to default value.
We get tables for unique and non-unique matches:
knitr::kable(fragmentMatchingPerformance(synobj2,
what="unique"))
ncommon | tp | fp | tn | fn | all | fdr |
---|---|---|---|---|---|---|
0 | 3097 | 79 | 0 | 0 | 4159 | 0.0248741 |
1 | 2556 | 22 | 57 | 541 | 2864 | 0.0085337 |
2 | 1682 | 12 | 67 | 1415 | 1864 | 0.0070838 |
3 | 1193 | 4 | 75 | 1904 | 1339 | 0.0033417 |
4 | 862 | 1 | 78 | 2235 | 987 | 0.0011587 |
5 | 651 | 1 | 78 | 2446 | 754 | 0.0015337 |
6 | 488 | 1 | 78 | 2609 | 578 | 0.0020450 |
7 | 372 | 1 | 78 | 2725 | 453 | 0.0026810 |
8 | 294 | 0 | 79 | 2803 | 367 | 0.0000000 |
9 | 240 | 0 | 79 | 2857 | 307 | 0.0000000 |
10 | 186 | 0 | 79 | 2911 | 245 | 0.0000000 |
11 | 149 | 0 | 79 | 2948 | 197 | 0.0000000 |
12 | 124 | 0 | 79 | 2973 | 165 | 0.0000000 |
13 | 95 | 0 | 79 | 3002 | 126 | 0.0000000 |
14 | 82 | 0 | 79 | 3015 | 105 | 0.0000000 |
15 | 59 | 0 | 79 | 3038 | 76 | 0.0000000 |
16 | 48 | 0 | 79 | 3049 | 61 | 0.0000000 |
17 | 41 | 0 | 79 | 3056 | 53 | 0.0000000 |
18 | 27 | 0 | 79 | 3070 | 36 | 0.0000000 |
19 | 21 | 0 | 79 | 3076 | 26 | 0.0000000 |
20 | 14 | 0 | 79 | 3083 | 16 | 0.0000000 |
21 | 12 | 0 | 79 | 3085 | 14 | 0.0000000 |
22 | 8 | 0 | 79 | 3089 | 10 | 0.0000000 |
23 | 6 | 0 | 79 | 3091 | 8 | 0.0000000 |
24 | 5 | 0 | 79 | 3092 | 6 | 0.0000000 |
25 | 4 | 0 | 79 | 3093 | 5 | 0.0000000 |
26 | 3 | 0 | 79 | 3094 | 3 | 0.0000000 |
27 | 3 | 0 | 79 | 3094 | 3 | 0.0000000 |
28 | 1 | 0 | 79 | 3096 | 1 | 0.0000000 |
knitr::kable(fragmentMatchingPerformance(synobj2,
what="non-unique"))
deltacommon | tp | fp | tn | fn | all | fdr |
---|---|---|---|---|---|---|
0 | 163 | 33 | 0 | 0 | 4159 | 0.1683673 |
1 | 163 | 2 | 31 | 0 | 374 | 0.0121212 |
2 | 132 | 1 | 32 | 31 | 304 | 0.0075188 |
3 | 119 | 1 | 32 | 44 | 276 | 0.0083333 |
4 | 110 | 0 | 33 | 53 | 256 | 0.0000000 |
5 | 95 | 0 | 33 | 68 | 226 | 0.0000000 |
6 | 85 | 0 | 33 | 78 | 202 | 0.0000000 |
7 | 75 | 0 | 33 | 88 | 178 | 0.0000000 |
8 | 67 | 0 | 33 | 96 | 161 | 0.0000000 |
9 | 60 | 0 | 33 | 103 | 145 | 0.0000000 |
10 | 53 | 0 | 33 | 110 | 127 | 0.0000000 |
11 | 43 | 0 | 33 | 120 | 106 | 0.0000000 |
12 | 37 | 0 | 33 | 126 | 92 | 0.0000000 |
13 | 27 | 0 | 33 | 136 | 71 | 0.0000000 |
14 | 20 | 0 | 33 | 143 | 56 | 0.0000000 |
15 | 12 | 0 | 33 | 151 | 37 | 0.0000000 |
16 | 10 | 0 | 33 | 153 | 26 | 0.0000000 |
17 | 8 | 0 | 33 | 155 | 21 | 0.0000000 |
18 | 7 | 0 | 33 | 156 | 18 | 0.0000000 |
19 | 4 | 0 | 33 | 159 | 9 | 0.0000000 |
20 | 2 | 0 | 33 | 161 | 5 | 0.0000000 |
21 | 2 | 0 | 33 | 161 | 5 | 0.0000000 |
22 | 2 | 0 | 33 | 161 | 5 | 0.0000000 |
23 | 1 | 0 | 33 | 162 | 3 | 0.0000000 |
24 | 1 | 0 | 33 | 162 | 3 | 0.0000000 |
25 | 1 | 0 | 33 | 162 | 3 | 0.0000000 |
Subsequently we could filter by minimal accepted common peaks:
filterUniqueMatches(synobj2, minNumber=1)
filterNonUniqueMatches(synobj2, minDelta=2)
filterNonUniqueIdentMatches(synobj2)
Finally we rescue EMRTs that are filtered but were identified by PLGS:
rescueEMRTs(synobj2, method="rescue")
In a similar manner as correcting for the retention time drift we correct
systematic errors of the intensity via a
LOWESS model.
The function modelIntensity
has to applied after findEMRTs
. The model is
build on the merged peptides as it is done for the retention time model. But in
contrast to the retention time model the prediction is necessary for the matched
quantitation data.
plotIntensity(synobj2, what="data")
setLowessSpan(synobj2, 0.05)
modelIntensity(synobj2)
plotIntensity(synobj2, what="model", nsd=1)
synergise2
The whole workflow described in the step-by-step workflow is
wrapped in the synergise2
function. As side effect it generates a nice
HTML report. An example could be found on https://github.com/lgatto/synapter.
synobj2 <- synergise2(filenames = inlist,
outputdir = ".")
For the next steps we need to convert the Synapter
object into an MSnSet
.
msn <- as(synobj2, "MSnSet")
Subsequently we look for synapter/PLGS agreement (this is more
useful for a combined MSnSet
; see basic synapter
online
or with vignette("synapter", package = "synapter")
).
synapterPlgsAgreement
adds an agreement column for each sample and counts the
agreement/disagreement in additional columns:
msn <- synapterPlgsAgreement(msn)
knitr::kable(head(fData(msn)[, grepl("[Aa]gree",
fvarLabels(msn)),
drop=FALSE]))
synapterPlgsAgreement.fermentor_02_sample_01_HDMSE_01 | nAgree | nDisagree | synapterPlgsAgreementRatio | |
---|---|---|---|---|
NVNDVIAPAFVK | agree | 1 | 0 | 1 |
TAGIQIVADDLTVTNPK | agree | 1 | 0 | 1 |
WLTGPQLADLYHSLMK | agree | 1 | 0 | 1 |
AAQDSFAAGWGVMVSHR | agree | 1 | 0 | 1 |
TFAEALR | agree | 1 | 0 | 1 |
LGANAILGVSLAASR | no_synapter_transfer | 0 | 0 | NaN |
As described in (Shliaha et al. 2013) Synapt G2 devices suffer from detector
saturation. This could be partly corrected by requantify
. Therefore a
saturationThreshold
has to be given above that intensity saturation
potentially happens. There are several methods available.
msncor <- requantify(msn,
saturationThreshold=1e5,
method="sum")
If an MSnSet
object was requantified using the "sum"
requantification method TOP3 normalisation is not valid anymore
because the most abundant proteins are penalised by removing high intensity
isotopes (for details see ?requantify
and ?rescaleForTop3
). This could be
overcome by calling rescaleForTop3
:
msncor <- rescaleForTop3(before=msn,
after=msncor,
saturationThreshold=1e5)
Since synapter 2.0 makeMaster
supports fragment files as well.
It is possible to create a fragment library that could used for
fragment matching because of the large data this could not
covered in this vignette. An introduction how to create a master could be
found in the basic synapter vignette, available
online
or with vignette("synapter", package = "synapter")
. Please find details
about creating a fragment library in ?makeMaster
.
All software and respective versions used to produce this document are listed below.
sessionInfo()
## R version 4.2.1 (2022-06-23)
## Platform: x86_64-pc-linux-gnu (64-bit)
## Running under: Ubuntu 20.04.5 LTS
##
## Matrix products: default
## BLAS: /home/biocbuild/bbs-3.15-bioc/R/lib/libRblas.so
## LAPACK: /home/biocbuild/bbs-3.15-bioc/R/lib/libRlapack.so
##
## locale:
## [1] LC_CTYPE=en_US.UTF-8 LC_NUMERIC=C
## [3] LC_TIME=en_GB LC_COLLATE=C
## [5] LC_MONETARY=en_US.UTF-8 LC_MESSAGES=en_US.UTF-8
## [7] LC_PAPER=en_US.UTF-8 LC_NAME=C
## [9] LC_ADDRESS=C LC_TELEPHONE=C
## [11] LC_MEASUREMENT=en_US.UTF-8 LC_IDENTIFICATION=C
##
## attached base packages:
## [1] stats4 stats graphics grDevices utils datasets methods
## [8] base
##
## other attached packages:
## [1] qvalue_2.28.0 BiocStyle_2.24.0 synapterdata_1.34.0
## [4] synapter_2.20.1 MSnbase_2.22.0 ProtGenerics_1.28.0
## [7] S4Vectors_0.34.0 mzR_2.30.0 Rcpp_1.0.9
## [10] Biobase_2.56.0 BiocGenerics_0.42.0
##
## loaded via a namespace (and not attached):
## [1] bitops_1.0-7 bit64_4.0.5 doParallel_1.0.17
## [4] RColorBrewer_1.1-3 GenomeInfoDb_1.32.4 tools_4.2.1
## [7] bslib_0.4.0 utf8_1.2.2 R6_2.5.1
## [10] affyio_1.66.0 cleaver_1.34.1 DBI_1.1.3
## [13] colorspace_2.0-3 tidyselect_1.1.2 bit_4.0.4
## [16] compiler_4.2.1 preprocessCore_1.58.0 cli_3.4.1
## [19] bookdown_0.29 sass_0.4.2 scales_1.2.1
## [22] readr_2.1.2 affy_1.74.0 stringr_1.4.1
## [25] digest_0.6.29 rmarkdown_2.16 XVector_0.36.0
## [28] pkgconfig_2.0.3 htmltools_0.5.3 highr_0.9
## [31] fastmap_1.1.0 limma_3.52.4 rlang_1.0.6
## [34] impute_1.70.0 jquerylib_0.1.4 generics_0.1.3
## [37] jsonlite_1.8.0 vroom_1.5.7 mzID_1.34.0
## [40] BiocParallel_1.30.3 dplyr_1.0.10 RCurl_1.98-1.8
## [43] magrittr_2.0.3 GenomeInfoDbData_1.2.8 MALDIquant_1.21
## [46] Matrix_1.5-1 munsell_0.5.0 fansi_1.0.3
## [49] MsCoreUtils_1.8.0 lifecycle_1.0.2 vsn_3.64.0
## [52] stringi_1.7.8 yaml_2.3.5 MASS_7.3-58.1
## [55] zlibbioc_1.42.0 plyr_1.8.7 grid_4.2.1
## [58] parallel_4.2.1 crayon_1.5.1 lattice_0.20-45
## [61] Biostrings_2.64.1 splines_4.2.1 multtest_2.52.0
## [64] hms_1.1.2 magick_2.7.3 knitr_1.40
## [67] pillar_1.8.1 reshape2_1.4.4 codetools_0.2-18
## [70] XML_3.99-0.10 glue_1.6.2 evaluate_0.16
## [73] pcaMethods_1.88.0 BiocManager_1.30.18 tzdb_0.3.0
## [76] vctrs_0.4.1 foreach_1.5.2 gtable_0.3.1
## [79] purrr_0.3.4 clue_0.3-61 assertthat_0.2.1
## [82] cachem_1.0.6 ggplot2_3.3.6 xfun_0.33
## [85] ncdf4_1.19 survival_3.4-0 tibble_3.1.8
## [88] iterators_1.0.14 IRanges_2.30.1 cluster_2.1.4
## [91] ellipsis_0.3.2
Gatto, Laurent, and Kathryn S Lilley. 2012. “MSnbase – an R/Bioconductor Package for Isobaric Tagged Mass Spectrometry Data Visualization, Processing and Quantitation.” Bioinformatics 28 (2): 288–9. https://doi.org/10.1093/bioinformatics/btr645.
Shliaha, Pavel Vyacheslavovich, Nicholas James Bond, Laurent Gatto, and Kathryn S. Lilley. 2013. “The Effects of Travelling Wave Ion Mobility Separation on Data Independent Acquisition in Proteomics Studies.” Journal of Proteome Research 2 (6). https://doi.org/10.1021/pr300775k.