An evolving computational platform for biological mass ... - PeerJ

10 downloads 1713 Views 3MB Size Report
Nov 17, 2015 - teapot for editing large data sets and a command line tool for transposing .... Different strategies permit the quantification of features: Label-free ...
An evolving computational platform for biological mass spectrometry: workflows, statistics and data mining with MASSyPup64 Robert Winkler Department of Biotechnology and Biochemistry, CINVESTAV Unidad Irapuato, Mexico

ABSTRACT

Submitted 7 September 2015 Accepted 22 October 2015 Published 17 November 2015 Corresponding author Robert Winkler, [email protected] Academic editor Rudiger Ettrich Additional Information and Declarations can be found on page 28 DOI 10.7717/peerj.1401 Copyright 2015 Winkler Distributed under Creative Commons CC-BY 4.0

In biological mass spectrometry, crude instrumental data need to be converted into meaningful theoretical models. Several data processing and data evaluation steps are required to come to the final results. These operations are often difficult to reproduce, because of too specific computing platforms. This effect, known as ‘workflow decay’, can be diminished by using a standardized informatic infrastructure. Thus, we compiled an integrated platform, which contains ready-to-use tools and workflows for mass spectrometry data analysis. Apart from general unit operations, such as peak picking and identification of proteins and metabolites, we put a strong emphasis on the statistical validation of results and Data Mining. MASSyPup64 includes e.g., the OpenMS/TOPPAS framework, the Trans-Proteomic-Pipeline programs, the ProteoWizard tools, X!Tandem, Comet and SpiderMass. The statistical computing language R is installed with packages for MS data analyses, such as XCMS/metaXCMS and MetabR. The R package Rattle provides a user-friendly access to multiple Data Mining methods. Further, we added the non-conventional spreadsheet program teapot for editing large data sets and a command line tool for transposing large matrices. Individual programs, console commands and modules can be integrated using the Workflow Management System (WMS) taverna. We explain the useful combination of the tools by practical examples: (1) A workflow for protein identification and validation, with subsequent Association Analysis of peptides, (2) Cluster analysis and Data Mining in targeted Metabolomics, and (3) Raw data processing, Data Mining and identification of metabolites in untargeted Metabolomics. Association Analyses reveal relationships between variables across different sample sets. We present its application for finding co-occurring peptides, which can be used for target proteomics, the discovery of alternative biomarkers and protein–protein interactions. Data Mining derived models displayed a higher robustness and accuracy for classifying sample groups in targeted Metabolomics than cluster analyses. Random Forest models do not only provide predictive models, which can be deployed for new data sets, but also the variable importance. We demonstrate that the later is especially useful for tracking down significant signals and affected pathways in untargeted Metabolomics. Thus, Random Forest modeling supports the unbiased search for relevant biological features in Metabolomics. Our results clearly manifest the importance of Data Mining methods to disclose non-obvious information in biological mass spectrometry . The application of a Workflow Management System

OPEN ACCESS

How to cite this article Winkler (2015), An evolving computational platform for biological mass spectrometry: workflows, statistics and data mining with MASSyPup64. PeerJ 3:e1401; DOI 10.7717/peerj.1401

and the integration of all required programs and data in a consistent platform makes the presented data analyses strategies reproducible for non-expert users. The simple remastering process and the Open Source licenses of MASSyPup64 (http://www. bioprocess.org/massypup/) enable the continuous improvement of the system. Subjects Bioinformatics, Biophysics, Computational Biology, Mathematical Biology, Statistics Keywords Bioinformatics, Computational mass spectrometry, Workflow decay, Metabolomics,

Workflow management systems, Data Mining, Proteomics, Model building, Association analyses, Random forest trees

INTRODUCTION Mass spectrometry provides qualitative and quantitative data about molecules. Since complex mixtures can be analyzed with high sensitivity and selectivity, mass spectrometry plays a central role in high-throughput biology (Jemal, 2000; Nilsson et al., 2010). Sequencing technologies have revolutionized the so-called ‘-omics’ sciences on the level of nucleic acids, genomics and transcriptomics (Sanger & Coulson, 1975; Wang, Gerstein & Snyder, 2009). But the study of the actual state of proteins and metabolites, which reflect the physiological condition of an organism, still relies mainly on mass spectrometry data. In proteomics, a combination of biochemical and instrumental techniques is used to obtain comprehensive, quantitative information about the expression, modification and degradation of proteins at a certain physiological state (Wilkins et al., 1996; Anderson & Anderson, 1998). Although gel electrophoresis, immuno-precipitation and other separation strategies are used as first focusing steps, the identification of proteins usually relies on mass spectrometry methods (Shevchenko et al., 2006). Metabolomics refers to the inventory of metabolites of an organism or tissue. The metabolome may be seen as an endpoint (Ernest et al., 2012), which derives from biochemical processes that depend on genomic and environmental factors. Therefore, the study of metabolic phenotypes allows both, the accurate classification of genotypes (Montero-Vargas et al., 2013; McClure, Chavarria & Espinoza, 2015; Musah et al., 2015) as well as an evaluation of the physiological state of an organism (Garc´ıa-Flores et al., 2012; Garc´ıa-Flores et al., 2015).

General mass spectrometry data processing workflow The data analysis of mass spectrometry experiments all follow the same logic, although the composition of the samples, the analytical question and the data format and quality might vary. A general workflow in biological mass spectrometry is given in Fig. 1 and consists of the following steps: Raw data import First of all, the raw data need to be converted into a format which is readable for the following data analysis programs. This step is not trivial, since the different manufacturers of mass spectrometers use a variety of proprietary data formats. Currently, the recommended

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

2/34

Figure 1 Universal workflow of mass spectrometry data analyses. Raw data need to be processed to extract informative features and to create statistically valid models.

standard by the Human Proteome Organization (HUPO) Proteomics Standards Initiative working group for mass spectrometry standards (PSI-MS) is mzML (Martens et al., 2011). Therefore, most MS data analysis programs are able to read and process this format. The ProteoWizard tools (http://proteowizard.sourceforge.net) allow the conversion of vendor-specific files to mzML archives (Chambers et al., 2012; Kessner et al., 2008). Since format-specific libraries are required, it is recommend to execute the conversion to mzML files directly on the control computer of the mass spectrometer. Alternatively, the ProteoWizard software can be installed with the vendor-libraries on a Windows computer. The ProteoWizard tools without licensed and Windows-specific libraries are available on MASSyPup64 for further pre-processing of MS data files. Spectra processing Spectra are collected either in ‘profile’ mode or in ‘centroid’ mode. Profile spectra still contain the shape of peaks and thus may provide additional information about the measured compounds. However, the size of the data archives might be considerable, especially for high resolution measurements. In contrast, centroid spectra only consist of mass-to-charge (m/z) values and their intensity. In many cases, it is advised to convert profile spectra to centroid spectra, to reduce computing effort.

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

3/34

Typical operations of spectra processing include a baseline substraction, smoothing, normalization, and peak picking. On MASSyPup64, various programs are available for these tasks, such as: msconvert (Chambers et al., 2012), OpenMS/TOPPAS (Sturm et al., 2008) and R/MALDIquant (Gibb & Strimmer, 2012). Some MS programs, such as Comet (Eng et al., 2015; Eng, Jahan & Hoopmann, 2013), X!Tandem (Craig & Beavis, 2004) and XCMS (Benton et al., 2008; Smith et al., 2006) do not require a prior external spectra processing, but can use raw mzML data as input. Feature analyses The mass spectrometry signals need to be transformed into chemical information. Therefore, ‘features’ have to be identified, which are e.g., defined by their m/z value and retention time. Usually the features display certain variations between samples, due to measurement tolerances. Those are corrected by an alignment of the feature maps, which finally allows to compare the abundance of features in different samples. Different strategies permit the quantification of features: Label-free quantification, the evaluation of different ion transitions (fragments of a molecule in so-called MultipleReaction-Monitoring, MRM) or the use of defined tags. The identification of molecules is desirable for most bioanalytical projects. For the identification of peptides and proteins, various search programs are available, which can be used or separately or in combination (Shteynberg et al., 2013). Identifying metabolites is still more challenging, although various databases, such as MassBank (http://www. massbank.jp/, (Horai et al., 2010)) and METLIN (https://metlin.scripps.edu/, (Smith et al., 2005)) as well as search algorithms have been published. The de-novo determination of chemical formulas from MS data is difficult, even with data from high-resolution instruments (Kind & Fiehn, 2006). Kind & Fiehn (2007) presented Seven Golden Rules (7GR) for the heuristic filtering of possible chemical formulas. The 7GR software was recently re-implemented for better usability and enriched with several functions. The respective program SpiderMass enables the construction of a custom data base with expected compounds for a certain biological context, which increases the probability of correctly identified metabolites (Winkler, 2015). Statistics and Data Mining Biological systems often exhibit notable variances, also measurement errors and wrong assignments of molecules are possible. Thus, usually biological and technical replicates are analyzed and the results are subjected to statistical analyses. More recently, Data Mining strategies are employed to unveil non-obvious information. Different approaches for Statistics and Data Mining are presented below, as well as their practical application to proteomics and metabolomics data sets. Integration and interpretation In a last step, the information obtained has to be interpreted within a biological context. Changes of protein concentrations can indicate the involvement of physiological processes. Metabolic information can lead to information about pathways which are affected in

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

4/34

Figure 2 Building of descriptive and predictive models. Data can be transformed either into descriptive models, which integrate all available data, or into predictive models. For generating predictive models, only part of the data is used to train the model, another part to validate the model, and finally the model can be tested with the resting data to estimate error rates.

certain conditions. Often, the identification of marker molecules is pursued, with the purpose to employ them later, e.g., for the early detection of diseases.

Statistics and Data Mining The American Statistical Association describes Statistics as “the science of learning from data, and of measuring, controlling, and communicating uncertainty; and it thereby provides the navigation essential for controlling the course of scientific and societal advances” (http://www.amstat.org/, Davidian & Louis (2012)). Accepting this broad definition, Data Mining (DM) is a sub-discipline of Statistics. Data Mining enhances ‘classic’ Statistics methods with machine learning (‘artificial intelligence’) algorithms and computer science. Data Mining supports the understanding of complex systems, which contain wealth of data with interacting variables. An important aspect of DM is the development of models, which represent the data in a structured form and support the extraction of information and creation of knowledge (Williams, 1987; Williams, 1988; Williams, 2011). Creation of models can be distinguished into descriptive and predictive (Fig. 2).

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

5/34

Descriptive models Descriptive models analyze relationships between variables or between individual samples. Since these models search for structures in a given data set, they are developed using the whole data set. Two important strategies are: • Association Analysis⋆ • Cluster Analysis⋆ Predictive models Predictive models search for rules, which connect input and output variables. Those variables can be categorical (tissue type, color, disease/healthy) or numeric. If the target is categorical, the final model performs a Classification. If the target is numeric, a Regression. Important model builders are: • Decision Tree Model⋆ • Random Forest Model⋆ • Support Vector Maschine (SVM) Model⋆ • Boost Model • Linear Regression Model⋆ • Neuronal Network Model For models marked with a ⋆ , a practical example in proteomics and/or metabolomics is given below. For more details about the knowledge representation of DM models, their algorithms and examples we refer to Williams (2011). Data Mining process and model development Data Mining (DM) is mostly used in Economics, e.g., for managing risks of bank loans or for detecting fraudulent activities. However, the activities for developing a model is similar for any DM project. The Cross Industry Standard Process for Data Mining (CRISP-DM) defines six phases (Shearer, 2000): 1. Business Understanding 2. Data Understanding 3. Data Preparation 4. Modeling 5. Evaluation 6. Deployment Obviously, in case of an Omics project we would replace ‘Business Understanding’ by ‘Problem Understanding’ or ‘Biological System Understanding’. The ‘Data Preparation’ is an important issue for analyzing mass spectrometry data. Depending on the number of samples and data quality, it might be necessary to eliminate variables or samples from the data set, to scale the data, to impute missing data points, etc. (Williams, 2011).

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

6/34

There is an important difference in the development of descriptive and predictive models. For descriptive models, the complete data set is used. For predictive models, the data set is separated into a training, a validation and a testing dataset, e.g., in a proportion 70:15:15 (Fig. 2). The training data serve for developing the model, the validation data set for monitoring the actual performance of the model, and the testing data for estimating the final performance of the model. Final models can be exchanged between different computing environments using the XML based Predictive Model Markup Language (PMML) format (Grossman, Hornick & Meyer, 2002). State of the art for statistics and DM in biological mass spectrometry For proteomics, bioinformatic pipelines are already well established. The different peptide/ protein search engines deliver distinct scores, which indicate the confidence of a identification hit, such as the Mascot score, the e-value or the XCorr (Kapp et al., 2005; Becker & Bern, 2011). But independently of the employed MS/MS search program, a subsequent statistical anaysis is necessary. The PeptideProphet and ProteinProphet algorithms allow the statistical modeling of peptide and protein identification results (Nesvizhskii et al., 2003; Keller et al., 2002). Using target-decoy database searches permit the estimation of false discovery rates (Elias & Gygi, 2007). Commercial, as well as Open Source platforms, integrate those individual programs to create complete proteomic workflows (Nelson et al., 2011; Keller et al., 2005; Rauch et al., 2006; Deutsch et al., 2010; Deutsch et al., 2015). Finally, the submission of results in standard formats to public databases makes the data available to the community (Griss & Gerner, 2009; Barsnes et al., 2009; Vizca´ıno, Foster & Martens, 2010; Cˆot´e et al., 2012; Vizca´ıno et al., 2013; Mohammed et al., 2014; Reisinger et al., 2015; Killcoyne, Deutsch & Boyle, 2012; Desiere et al., 2006). In metabolomics, still more issues are awaiting resolution. E.g., the unequivocal assignment of mass signals to the correct compounds and the estimation of the statistical confidence of metabolite identifications is still challenging. The R packages XCMS/XCMS2 (Smith et al., 2006; Benton et al., 2008) and metaXCMS (Tautenhahn et al., 2011; Patti, Tautenhahn & Siuzdak, 2012) permit the realization of complete metabolic workflows and the comparison of various samples. Correct application of included functions improve the detection, quantification and identification of metabolites (METLIN database, (Benton, Want & Ebbels, 2010; Tautenhahn, B¨ottcher & Neumann, 2008; Smith et al., 2005)). The XCMS collection is technically mature and comprehensive, but for most casual users too complicated to handle. XCMS Online (Tautenhahn et al., 2012) facilitates the use of XCMS by non-experts. However, the control over data and the option to optimize the code for project-specific needs is limited in the online version. MZmine 2 is another, java-based, framework for mass spectrometry data workflows with some statistical tools such as Principle Components Analysis (PCA) and Clustering capabilities, which is especially user-friendly and extensible (Pluskal et al., 2010).

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

7/34

Resuming, various bioinformatic solutions are already available for the processing and statistical analysis of proteomics and metabolomics data. But the concept of Data Mining is still not implemented in current biological mass spectrometry. The traditional Omics approach is exploratory and starts from a biological question or problem. Usually it is rather curiosity- than hypothesis-driven. An Omics study normally ends with a statistically valid descriptive model, which is interpreted from a biological point of view. Often, the results help to build theories or hypotheses, which are testable afterwards. In stark contrast, predictive models from Data Mining modeling can be immediately deployed and support decision making. Especially clinical applications (biomarker studies) and projects with limited sample availability (ecology, identification of microorganisms, ‘Biotyping’) could greatly benefit from the implementation of Data Mining strategies. Data Mining algorithms are also capable to uncover rules or patterns in complex data structures, without being biased by a (bio)scientist’s expectations.

Aim of this study Data Mining strategies promise high potential for the analysis of biological mass spectrometry data, but there is still scarce use of it in current MS based Omics studies. On the other side, there is already a rich variety of excellent software for mass spectrometry data processing software (http://www.ms-utils.org/), and also for statistics and Data Mining available (Williams, 2011; Gibb & Strimmer, 2012; Luca Belmonte & Nicolini, 2013; Williams, 2009). Thus, we compiled a computational platform for the high-throughput data analysis in proteomics and metabolomics, which facilitates the rapid set-up of workflows and the subsequent Data Mining. MASSyPup64 (http://www.bioprocess.org/massypup) is a 64-bit live system, which can be run directly from external media. Open Source licenses of the software and the remastering utility provided on Fatdog64 promote the further development and the adjustment to the needs of a laboratory. Based on real datasets from proteomics and targeted and untargeted metabolomics we demonstrate the creation of efficient data processing workflows. Further, we stress out the opportunity to discover non-obvious biological knowledge by Data Mining methods in biological mass spectrometry.

METHODS Operating system The original MASSyPup distribution was built on a 32-bit platform and contains multiple programs for analyzing mass spectrometry data Winkler (2014). The new MASSyPup64 is much more focused on the high-throughput processing of ‘big data’ and the subsequent Data Mining. MASSyPup64 is bootable on Windows (including with EFI ‘secure boot’) and Linux PCs.

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

8/34

As starting point, the 64-bit Linux distribution Fatdog64 was chosen (http://distro. ibiblio.org/fatdog/web/). The system is preferably installed on DVD or USB media. Data from all available local drives are accessible for analysis. The mass spectrometry programs, special libraries and additional tools were compiled, if necessary and installed in the directory branch of /usr/local. Most programs can be started directly from a console window. For Python (https://www.python.org/), versions 2 and 3 are installed. The default Python 2 is called by python, version 3 by python3. Fatdog64 contains already a remastering tool with Graphical User Interface. Since the MASSyPup64 version already occupies several Gigabytes, it is recommendable to choose the “small initrd” option. The current release of MASSyPup64, as well as FNAs (Frequently Needed Answers) and a list of currently installed software can be found on the project homepage (http://www. bioprocess.org/massypup/). All components are Open Source software, which permits the free distribution and modification of the system.

Workflow management systems The ideal Workflow Management System (WMS) should be visual, modular and easy to understand. The facile integration of external commands and the development of new functions should be possible. Further, the WMS should allow to analyze data, which are stored outside the running platform, i.e., without uploading the data to the WMS. The last requirement is important, since mass spectrometry projects often exceed various Gigabytes of data volume and thus copying or moving them is inconvenient. After trying various options, two WMS were installed on MASSyPup64: 1. TOPPAS 2. Taverna The Trans-Proteomic-Pipeline 4.8.0 was compiled and installed on MASSyPup64, but without configuring the hosting server. Consequently, the TPP tools are available for being employed in workflows, but the web-interface is not running. Below, a workflow emulating the TPP for protein identification and validation is demonstrated.

Statistics and Data Mining For statistical analyses, Data Mining and graphics, we compiled and installed an ‘R’ software environment (https://www.r-project.org/). A large scientific community is contributing to this powerful programming language (The R Journal, http://journal. r-project.org/). Above mentioned XCMS/XCMS2/metaXCMS (Smith et al., 2006; Benton et al., 2008; Tautenhahn et al., 2011; Patti, Tautenhahn & Siuzdak, 2012) packages were installed, as well as MALDIquant/MALDIquantForeign for spectra processing (Gibb & Strimmer, 2012) and MSI.R for evaluating Mass Spectrometry Imaging (MSI) data (Gamboa-Becerra et al., 2015).

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

9/34

For the linear model analysis of metabolomic data, we included MetabR (http://metabr. r-forge.r-project.org/), which provides a Graphical User Interface (GUI) and can be used for both, statistical data evaluation and data preparation for Mining. Rattle—the R Analytical Tool To Learn Easily—represents a sophisticated and free environment for Data Mining (http://rattle.togaware.com/, Williams, 2011; Williams, 2009). The GUI facilitates the loading, visualization and exploration of data, especially for beginners without profound R knowledge. Rattle also supports the export of models in PMML standard format and was installed on MASSyPup64 with all suggested packages (including database connectors, ggobi http://www.ggobi.org/, etc.).

Special tools for large data set editing and shaping Standard spreadsheet software such as Excel, LibreOffice Calc or Gnumeric become very sluggish, if it comes to the editing of large tables. With R, huge tables can be handled and various of GNU programs (http://www.gnu.org/manual/blurbs.html), such as grep, sed, wc, .. can be used to edit big data files. But the import and manipulation of data is not always very practical with those tools. Therefore, some special programs for data manipulation were included into the MASSyPup64 distribution. Spreadsheet program teapot The non-traditional ‘Table Editor And Planner, Or: Teapot!’ was originally developed by Michael Haardt and J¨org Walter and is currently hosted at SYNTAX-K http://www. syntax-k.de/projekte/teapot/. For best performance and usability, Teapot was re-compiled and statically linked with the FLTK GUI toolkit (http://www.fltk.org). Large matrix transposing Frequently, it is necessary to transpose a data matrix before loading it into another program. This can be efficiently done with the command ‘transpose’ (version 2.0 by Dr. Alex Sheppard, http://www.das-computer.co.uk). The C program was modified and re-compiled to change the default maximum matrix size to 100,000 × 100,000.

RESULTS AND DISCUSSION Operating system and installed programs Based on the Linux platform Fatdog64 (http://distro.ibiblio.org/fatdog/web/), an analysis framework and programming environment for mass spectrometry data was created. Table 1 lists the programs currently installed on MASSyPup64 (standard software such as editor, browser etc. are included as well, but not explicitly mentioned in the table). For installation, the iso image has to be burned on a DVD or installed onto a USB stick (e.g., with Rufus from https://rufus.akeo.ie). Upgrading individual programs or libraries is not necessary, because each release is a complete compilation of operating system and software. User settings, programs and data may be stored on local devices and are kept for future releases. If additional programs are required, they can be installed and included into a new MASSyPup64 version by using the Fatdog64 remastering tool.

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

10/34

Table 1 Programs installed on MASSyPup64. Development tools and libraries devx jdk java Java(TM) SE Runtime Environment Java HotSpot(TM) 64-Bit Server VM Python Workflow management taverna OpenMS/TOPPAS Data conversion and trimming ProteoWizard tools

Development tools for fatdog64-701, such as C/C++, FORTRAN compiler, header libraries Java Development Toolkit Version 1.7.0 65 Build 1.7.0 65-b17 Build 24.65-b04, mixed mode Version 2 (python) and version 3 (python3) Java-based workflow management with additional functions (condition, iteration) Framework for mass spectrometry data processing and GUI supported design of pipelines MS data conversion and tools, see /usr/local/massypup64/proteowizard-tools.txt; without vendor libraries Transposing matrix data from commandline, source code in /usr/local/bin/transpose.c. Currently compiled for a default maximum matrix size of 10,000 × 10,000 Spreadsheet program supporting three dimensional data sets. The manual “teapot.pdf ” is located in /usr/local/doc

transpose

teapot/fteapot MS data processing suite MZmine Proteomics search engines Comet OMSSA X!Tandem TPP 4.8 R statistics language XCMS/metaXCMS Metab.R Rattle Compound identification massXpert SpiderMass Mass Spectrometry Imaging (MSI) HelloPhidget OpenMZxy MSI.R Data visualization GGobi

Java-based MS analysis program, focused on metabolomics MS/MS protein search algorithm MS/MS protein search algorithm MS/MS protein search algoritm Trans-Proteomic-Pipeline Metabolomics Statistical analysis of metabolomics results Data Mining Analysis of polymer spectra Target DB creation/ matching, Online searches and formula generation Test tool for detection of connected Phidgets (prototyping of MSI) Control of Phidget imaging robot MSI analysis with R scripts Interactive data visualization and exploration tool

Proteomics: identification of proteins, peptideprophet/proteinprophet validation, text mining and association analysis Data set and bioanalytical question Peroxidases are related to the post-harvest insect resistance of maize kernels (Winkler & Garc´ıa-Lara, 2010; Garc´ıa-Lara et al., 2007). Therefore, protein fractions of highly

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

11/34

insect-resistance maize kernels with peroxidase activity were subjected to 1D or 2D gel electrophoresis and subsequently analyzed with LC-MS/MS. The data set consists of three samples: 2DM, a spot from 2D gel electrophoresis of maize kernels with peroxidase activity; 1DM, a protein band from 1D gel electrophoresis of partially purified peroxidase, and 1DR, a protein band with peroxidase activity from recombinant production of a putative peroxidase, which was cloned from cDNA. Details about the project can be found at L´opez-Castillo et al. (2015). The workflow should identify potential candidates of peroxidases and suggest peptides for a targeted screening of peroxidases. Taverna Workflow Design The design of the proteomics workflow using taverna was inspired by the work of Bruin, Deelder & Palmblad (2012), but several modifications were undertaken: Peptide search. The search engine X!Tandem (Craig & Beavis, 2004) was replaced by Comet (Eng et al., 2015; Eng, Jahan & Hoopmann, 2013), in order to simplify the configuration by the user. All necessary parameters for the peptide identification are defined in the comet.params file, which has to be located in the same directory as the raw data files, which are expected in mzML format. A template for the comet.params file can be created by invoking the command comet -p. The location of the protein (or DNA) database is set with the database name option. For performing a concatenated decoy search (Elias & Gygi, 2007),the parameter decoy search needs to be set to 1. The separate generation of a decoy database is not required anymore. PeptideProphet/ProteinProphet validation. The results of the Comet search are written directly to pep.xml format and can be passed to the PeptideProphet script (Keller et al., 2002). The processed pep.xml files are subsequently evaluated using ProteinProphet (Nesvizhskii et al., 2003). Both validation programs are part of the Trans-Proteomic-Pipeline (TPP) (Keller et al., 2005; Deutsch et al., 2010; Deutsch et al., 2015) and integrated into the workflow by very simple tool modules, which facilitate the modification of parameters by advanced users. Creation of output in different formats. After the writing of the validated prot.xml files, the results are exported into various formats for further evaluation: The spreadsheet format xls (compatible with gnumeric and EXCEL), coma separated values (csv) text files, html (for opening the results in an internet browser) and in mzid (mzIdentML), a standard format for reporting proteomics results. The used tools were adopted from the TPP, the OpenMS/TOPPAS framework and from the Linux system programs, which underlines the flexibility of the taverna WMS. Text extraction. In the last module, protein hits, which contain the defined search pattern for proteins of interest, are written into a separate summary file in csv format. This simple Text Mining step allows the rapid screening for relevant identification results. An illustration of the complete workflow is given in Fig. 3.

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

12/34

Figure 3 Proteomics workflow with validation of hits by PeptideProphet/ ProteinProphet and final extraction of hits for proteins of potential interest.

For running the workflow (/usr/local/massypup64-taverna-workflows/m64-comet-valexport.t2flow), only the data path ‘/usr/local/massypup64-examples/ Maize-Proteomics-PODs’ needs to be given (as a value), and a string for the protein(s) of interest. In this example, “eroxidase” was defined as search string (omitting the initial letter P/p to avoid possible case problems). A version of the workflow without the extraction module is also stored in the workflow examples directory. This workflow can be used for a batch-wise protein identification similar to the Trans-Proteomic-Pipeline. The short workflow only requires the mzML data/comet.params directory as input value. Workflow results Running the workflow delivers the expected output files, as well as sensitivity vs. error plots for the individual samples (see Fig. 4). Table 2 corresponds to the exported hits of putative peroxidades. Considering a minimum of two unique peptides and a probability of at least 0.95 as acceptance criteria, no peroxidase (POD) related protein was identified in the 2D spot, five POD candidates in the purified fraction from the 1D gel, and two PODs from the 1D gel after recombinant production of the putative POD B6T173 in Escherichia coli. Thus, the workflow allows a rapid screening for proteins of interest. Indeed, further biochemical experiments confirmed protein B6T173 as the responsible one for the POD activity in the maize kernel. Association analysis Association Analysis investigates the probability of the co-occurrence of items. It is mainly known from Market Basket studies and social networks. For instance, if a person buys a

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

13/34

Figure 4 Plot of estimated sensitivity vs. error for sample 1DM, as delivered by the taverna workflow. Figure modified from original program output for improved readability.

telescope, most likely (s)he also might be interestd in a star map. Or if Henry knows Peter and Paul, he probably knows Mary as well. Importantly, the Association Analysis does not query the causality, but the likelihood of a relationship. Although the occurrence of an association might be low—let’s say the mentioned group of Henry, Peter, Paul and Mary represents a fraction of 0.0001% of a social network—the confidence might be high, e.g., 0.9, and therefore be highly informative. To search out co-occurring peptides, which could lead to possibly associated proteins and suitable peptides for targeted proteomics, we carried out an Association Analysis with Rattle. In total, more than 700 peptides with a probability above 0.9 were considered. A minimal support of 0.6, a minimal confidence of 0.9 and a path of seven rules were chosen as parameters. Fig. 5 represents the associations between seven peptides, which are related. Table 3 lists the associated peptides together with their identifications. DSACSAGGLEYEVPSGRR, TDPSVDPAYAGHLK, VQVLTGDEGEIR are genuine peptides of B6T173. The peptide TVSCADVLAFAAR is not present in the amio acid sequence of B6T173, but the similar peptide TVSCADIVAFAAR. Since B6T173 was recombinately produced in E. coli (sample 1DR), the identification of this peptide indicates an unexpected phenomenon during the MS measurement. However, the respective transitions might be useful for the quantification of the protein. The appearance of chloroplast protein might be feasible for the maize derived samples, but are unlikely to reflect reality in the bacterial preparation of B6T173. The glycine-rich peptide can be found in many organisms and therefore does not contribute information. Resuming, a set of 3+1 peptides was found, which are highly indicative for the protein B6T173. Since the protein is related to post-harvest insect resistance, those peptide MS

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

14/34

Table 2 Identified putative peroxidases, after PeptideProphet/ ProteinProphet validation. Sample

Accession

Protein Prob.

Coverage

Unique peps

Description

2DM 1DM 1DM 1DM 1DM 1DM 1DM 1DM 1DM 1DR 1DR 1DR 1DR 1DR 1DR 1DR 1DR 1DR 1DR 1DR 1DR 1DR 1DR

B4FBY8 B4FK72 B6T173 K7TID5 K7TID0 B4FY83 B4FNL8 B6SI04 K7VNV5 K7TID5 B6T173 B4FSW5 B4FY83 K7TMB4 Q6JAH6 K7V8K5 B4FNI0 A0A0B4J371 B4FBC8 B4G0X5 B6TWB1 C0PKS1 Q9ZTS6

0.6181 1.0000 1.0000 1.0000 0.9937 0.9890 0.0000 0.0000 0.0000 1.0000 0.9995 0.9990 0.9990 0.9990 0.6603 0.5743 0.3475 0.0000 0.0000 0.0000 0.0000 0.0000 0.0000

5.9 2.7 12.7 39.5 9.5 3.7

1 2 7 24 1 2 0 0 0 7 2 1 1 1 1 1 1 0 0 0 0 0 0

Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Glutathione peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase Peroxidase K (Fragment)

17.7 7.1 2.9 3.7 3.3 7.1 3.0 5.4

Table 3 Association Analysis of peptides across three samples. ‘x’ stands for peptides present in the sample and ‘o’ for peptides missing in the sample. Peptide

2DM

1DM

1DR

Acession

Description

DSACSAGGLEYEVPSGRR TDPSVDPAYAGHLK TVSCADVLAFAAR VQVLTGDEGEIR AFVHGDGDLFSR LFLNLQKEMNSVMVTRK GSGGGGGGGGGQGQSR

o o o o x o x

x x x x x x x

x x x x x x x

K7TID5 B6T173 B4FY83 K7TID5 B6SRJ2 A0A096PYN5 A0A096RDU5

Peroxidase Peroxidase Peroxidase Peroxidase Senescence-inducible chloroplast stay-green protein 30S ribosomal protein S2, chloroplastic Uncharacterized protein

transitions could serve for the screening of seeds. The PeptidePicker workflow delivers theoretical peptides for targeted proteomics (Mohammed et al., 2014). However, if experimental data are available, the Association Analysis includes all possible variables which affect the peptide/ protein identification from sample extraction to final evaluation, and thus should suggest more reliable candidates.

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

15/34

Graph for 7 rules

size: support (0.667 − 0.667) color: lift (1 − 1.5)

TVSCADVLAFAAR LFLNLQKEMNSVMVTRK

GSGGGGGGGGGQGQSR VQVLTGDEGEIR

AFVHGDGDLFSR

TDPSVDPAYAGHLK DSACSAGGLEYEVPSGRR

Figure 5 Associated peptides across the samples.

Ideally, an Association Analysis is carried out with numerous individual samples, which allows to reduce the support limit and to bring to light non-obvious correlations between variables or observations. Apart from finding reliable peptides for protein quantification, Association Analyses can be employed to discover alternative biomarkers, e.g., if the genuine biomarker is difficult to detect, or to search for protein–protein interactions.

Targeted metabolomics: cluster analyses, linear model analysis and model building using data mining Data set and analytical question We re-processed a data set which was described by Ernest et al. (2012). To study the adipose tissue metabolism, three groups of chicken where analyzed, which underwent different treatments: A control group (“Control”, sample 1–7) which were fed ad libitum, chicken fasted for 5 h (“Fast”, sample 8–14) and a group treated with an insulin inhibitor (“InsNeut”, sample 15–21). For more details about the biological experiments, we refer to the original paper Ernest et al. (2012). From the targeted metabolomics data, a statistical analysis yielding fold-changes and p-values should be carried out. Further, a classification of the three groups, based to their metabolic profile, should be intended. Statistical evaluation with MetabR Using MetabR, the fold-differences and the Tukey’s Honest Significant Difference (HSD) was calculated, applying a fixed linear model for the variables “Quantity” and

Winkler (2015), PeerJ, DOI 10.7717/peerj.1401

16/34

Table 4 Statistical analysis of targeted metabolomics data with MetabR. Bold values are significant with p-values

Suggest Documents