Publications
Domingo-Calap, Pilar; Schubert, Benjamin; Joly, Mélanie; Solis, Morgane; Untrau, Meiggie; Carapito, Raphael; Georgel, Philippe; Caillard, Sophie; Fafi-Kremer, Samira; Paul, Nicodème; Kohlbacher, Oliver; González-Candelas, Fernando; Bahram, Seiamak An unusually high substitution rate in transplant-associated BK polyomavirus in vivo is further concentrated in HLA-C-bound viral peptides PLoS Pathogens, 14 (10), pp. e1007368, 2018. @article{PlosPathHLAC2018, title = {An unusually high substitution rate in transplant-associated BK polyomavirus in vivo is further concentrated in HLA-C-bound viral peptides}, author = {Pilar Domingo-Calap and Benjamin Schubert and Mélanie Joly and Morgane Solis and Meiggie Untrau and Raphael Carapito and Philippe Georgel and Sophie Caillard and Samira Fafi-Kremer and Nicodème Paul and Oliver Kohlbacher and Fernando González-Candelas and Seiamak Bahram}, url = {https://dx.doi.org/10.1371%2Fjournal.ppat.1007368}, year = {2018}, date = {2018-01-01}, journal = {PLoS Pathogens}, volume = {14}, number = {10}, pages = {e1007368}, abstract = {Infection with human BK polyomavirus, a small double-stranded DNA virus, potentially results in severe complications in immunocompromised patients. Here, we describe the in vivo variability and evolution of the BK polyomavirus by deep sequencing. Our data reveal the highest genomic evolutionary rate described in double-stranded DNA viruses, i.e., 10-3-10-5 substitutions per nucleotide site per year. High mutation rates in viruses allow their escape from immune surveillance and adaptation to new hosts. By combining mutational landscapes across viral genomes with in silico prediction of viral peptides, we demonstrate the presence of significantly more coding substitutions within predicted cognate HLA-C-bound viral peptides than outside. This finding suggests a role for HLA-C in antiviral immunity, perhaps through the action of killer cell immunoglobulin-like receptors. The present study provides a comprehensive view of viral evolution and immune escape in a DNA virus.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Infection with human BK polyomavirus, a small double-stranded DNA virus, potentially results in severe complications in immunocompromised patients. Here, we describe the in vivo variability and evolution of the BK polyomavirus by deep sequencing. Our data reveal the highest genomic evolutionary rate described in double-stranded DNA viruses, i.e., 10-3-10-5 substitutions per nucleotide site per year. High mutation rates in viruses allow their escape from immune surveillance and adaptation to new hosts. By combining mutational landscapes across viral genomes with in silico prediction of viral peptides, we demonstrate the presence of significantly more coding substitutions within predicted cognate HLA-C-bound viral peptides than outside. This finding suggests a role for HLA-C in antiviral immunity, perhaps through the action of killer cell immunoglobulin-like receptors. The present study provides a comprehensive view of viral evolution and immune escape in a DNA virus. |
Deutsch, Eric; Perez-Riverol, Yasset; Chalkley, Robert; Wilhelm, Mathias; Tate, Stephen; Sachsenberg, Timo; Walzer, Mathias; Käll, Lukas; Delanghe, Bernard; Böcker, Sebastian; Schymanski, Emma; Wilmes, Paul; Dorfer, Viktoria; Kuster, Bernhard; Volders, Pieter-Jan; Jehmlich, Nico; Vissers, Johannes; Wolan, Dennis; Wang, Ana; Mendoza, Luis; Shofstahl, Jim; Dowsey, Andrew; Griss, Johannes; Salek, Reza; Neumann, Steffen; Binz, Pierre-Alain; Lam, Henry; Vizcaíno, Juan; Bandeira, Nuno; Röst, Hannes Expanding the use of spectral libraries in proteomics Journal of Proteome Research, 2018. @article{JPRSpecLib2018, title = {Expanding the use of spectral libraries in proteomics}, author = {Eric Deutsch and Yasset Perez-Riverol and Robert Chalkley and Mathias Wilhelm and Stephen Tate and Timo Sachsenberg and Mathias Walzer and Lukas Käll and Bernard Delanghe and Sebastian Böcker and Emma Schymanski and Paul Wilmes and Viktoria Dorfer and Bernhard Kuster and Pieter-Jan Volders and Nico Jehmlich and Johannes Vissers and Dennis Wolan and Ana Wang and Luis Mendoza and Jim Shofstahl and Andrew Dowsey and Johannes Griss and Reza Salek and Steffen Neumann and Pierre-Alain Binz and Henry Lam and Juan Vizcaíno and Nuno Bandeira and Hannes Röst}, url = {https://pubs.acs.org/doi/10.1021/acs.jproteome.8b00485}, year = {2018}, date = {2018-01-01}, journal = {Journal of Proteome Research}, abstract = {The 2017 Dagstuhl Seminar on Computational Proteomics provided an opportunity for a broad discussion on the current state and future directions of the generation and use of peptide tandem mass spectrometry spectral libraries. Their use in proteomics is growing slowly, but there are multiple challenges in the field that must be addressed to further increase the adoption of spectral libraries and related techniques. The primary bottlenecks are the paucity of high quality and comprehensive libraries and the general difficulty of adopting spectral library searching into existing workflows. There are several existing spectral library formats, but none capture a satisfactory level of metadata; therefore a logical next improvement is to design a more advanced, Proteomics Standards Initiative-approved spectral library format that can encode all of the desired metadata. The group discussed a series of metadata requirements organized into three designations of completeness or quality, tentatively dubbed bronze, silver, and gold. The metadata can be organized at four different levels of granularity: at the collection (library) level, at the individual entry (peptide ion) level, at the peak (fragment ion) level, and at the peak annotation level. Strategies for encoding mass modifications in a consistent manner and the requirement for encoding high-quality and commonly-seen but as-yet-unidentified spectra were discussed. The group also discussed related topics, including strategies for comparing two spectra, techniques for generating representative spectra for a library, approaches for selection of optimal signature ions for targeted workflows, and issues surrounding the merging of two or more libraries into one. We present here a review of this field and the challenges that the community must address in order to accelerate the adoption of spectral libraries in routine analysis of proteomics datasets.}, keywords = {}, pubstate = {published}, tppubtype = {article} } The 2017 Dagstuhl Seminar on Computational Proteomics provided an opportunity for a broad discussion on the current state and future directions of the generation and use of peptide tandem mass spectrometry spectral libraries. Their use in proteomics is growing slowly, but there are multiple challenges in the field that must be addressed to further increase the adoption of spectral libraries and related techniques. The primary bottlenecks are the paucity of high quality and comprehensive libraries and the general difficulty of adopting spectral library searching into existing workflows. There are several existing spectral library formats, but none capture a satisfactory level of metadata; therefore a logical next improvement is to design a more advanced, Proteomics Standards Initiative-approved spectral library format that can encode all of the desired metadata. The group discussed a series of metadata requirements organized into three designations of completeness or quality, tentatively dubbed bronze, silver, and gold. The metadata can be organized at four different levels of granularity: at the collection (library) level, at the individual entry (peptide ion) level, at the peak (fragment ion) level, and at the peak annotation level. Strategies for encoding mass modifications in a consistent manner and the requirement for encoding high-quality and commonly-seen but as-yet-unidentified spectra were discussed. The group also discussed related topics, including strategies for comparing two spectra, techniques for generating representative spectra for a library, approaches for selection of optimal signature ions for targeted workflows, and issues surrounding the merging of two or more libraries into one. We present here a review of this field and the challenges that the community must address in order to accelerate the adoption of spectral libraries in routine analysis of proteomics datasets. |
Karim, Md. Rezaul; Nguyen, Binh-Phi; Zimmermann, Lukas; Kirsten, Toralf; Löbe, Matthias; Meineke, Frank; Stenzhorn, Holger; Kohlbacher, Oliver; Decker, Stefan; Beyan, Oya (Ed.) A Distributed Analytics Platform to Execute FHIR based Phenotyping Algorithms 2018. (BibTeX) @proceedings{PHTFHIR2018, title = {A Distributed Analytics Platform to Execute FHIR based Phenotyping Algorithms}, editor = {Md. Rezaul Karim and Binh-Phi Nguyen and Lukas Zimmermann and Toralf Kirsten and Matthias Löbe and Frank Meineke and Holger Stenzhorn and Oliver Kohlbacher and Stefan Decker and Oya Beyan}, year = {2018}, date = {2018-01-01}, series = {11th International Conference on Semantic Web Applications and Tools for Healthcare and Life Sciences (SWAT4HCLS'2018)}, keywords = {}, pubstate = {published}, tppubtype = {proceedings} } |
Wein, Samuel; Andrews, Byron; Sachsenberg, Timo; Santos-Rosa, Helena; Kohlbacher, Oliver; Kouzarides, Tony; Garcia, Benjamin A; Weisser, Hendrik A computational platform for high-throughput analysis of RNA sequences and modifications by mass spectrometry bioRxiv, 2018. @article{Wein501668, title = {A computational platform for high-throughput analysis of RNA sequences and modifications by mass spectrometry}, author = {Samuel Wein and Byron Andrews and Timo Sachsenberg and Helena Santos-Rosa and Oliver Kohlbacher and Tony Kouzarides and Benjamin A Garcia and Hendrik Weisser}, url = {https://www.biorxiv.org/content/early/2018/12/19/501668}, year = {2018}, date = {2018-01-01}, journal = {bioRxiv}, abstract = {The field of epitranscriptomics is growing in importance, with chemical modification of RNA being associated with a wide variety of biological phenomena. A pivotal challenge in this area is the identification of modified RNA residues within their sequence contexts. Next-generation sequencing approaches are generally unable to capture modifications, although workarounds for some epigenetic marks exist. Mass spectrometry (MS) offers a comprehensive solution by using analogous approaches to shotgun proteomics. However, software support for the analysis of RNA MS data is inadequate at present and does not allow high-throughput processing. In particular, existing software solutions lack the raw performance and statistical grounding to efficiently handle the large variety of modifications present on RNA. We present a free and open-source database search engine for RNA MS data, called NucleicAcidSearchEngine (NASE), that addresses these shortcomings. We demonstrate the capability of NASE to reliably identify a wide range of modified RNA sequences in three original datasets of varying complexity. In a human tRNA sample, we characterize over 20 different modification types simultaneously and find many cases of incomplete modification.}, keywords = {}, pubstate = {published}, tppubtype = {article} } The field of epitranscriptomics is growing in importance, with chemical modification of RNA being associated with a wide variety of biological phenomena. A pivotal challenge in this area is the identification of modified RNA residues within their sequence contexts. Next-generation sequencing approaches are generally unable to capture modifications, although workarounds for some epigenetic marks exist. Mass spectrometry (MS) offers a comprehensive solution by using analogous approaches to shotgun proteomics. However, software support for the analysis of RNA MS data is inadequate at present and does not allow high-throughput processing. In particular, existing software solutions lack the raw performance and statistical grounding to efficiently handle the large variety of modifications present on RNA. We present a free and open-source database search engine for RNA MS data, called NucleicAcidSearchEngine (NASE), that addresses these shortcomings. We demonstrate the capability of NASE to reliably identify a wide range of modified RNA sequences in three original datasets of varying complexity. In a human tRNA sample, we characterize over 20 different modification types simultaneously and find many cases of incomplete modification. |
Halfmann, Marc; Stenzhorn, Holger; Gerjets, Peter; Kohlbacher, Oliver; Oestermeier, Uwe User-Driven Development of a Novel Molecular Tumor Board Support Tool Vidal, Sören AuerMaria-Esther (Ed.): DILS 2018: Data Integration in the Life Sciences, pp. 195-199, Springer, 2018. @inproceedings{MTB_DILS2018, title = {User-Driven Development of a Novel Molecular Tumor Board Support Tool}, author = {Marc Halfmann and Holger Stenzhorn and Peter Gerjets and Oliver Kohlbacher and Uwe Oestermeier}, editor = {Sören AuerMaria-Esther Vidal}, url = {https://link.springer.com/chapter/10.1007/978-3-030-06016-9_18}, year = {2018}, date = {2018-01-01}, booktitle = {DILS 2018: Data Integration in the Life Sciences}, volume = {11371}, pages = {195-199}, publisher = {Springer}, series = {Lecture Notes in Computer Science}, abstract = {Nowadays personalized medicine is of increasing importance, especially in the field of cancer therapy. More and more hospitals are conducting molecular tumor boards (MTBs) bringing together experts from various fields with different expertise to discuss patient cases taking into account genetic information from sequencing data. Yet, there is still a lack of tools to support collaborative exploration and decision making. To fill this gap, we developed a novel user interface to support MTBs. A task analysis of MTBs currently held at German hospitals showed, that there is less collaborative exploration during the meeting as expected, with a large part of the information search being done during the MTB preparation. Thus we designed our interface to support both situations, a single user preparing the MTB and the presentation of information and group discussion during the meeting.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Nowadays personalized medicine is of increasing importance, especially in the field of cancer therapy. More and more hospitals are conducting molecular tumor boards (MTBs) bringing together experts from various fields with different expertise to discuss patient cases taking into account genetic information from sequencing data. Yet, there is still a lack of tools to support collaborative exploration and decision making. To fill this gap, we developed a novel user interface to support MTBs. A task analysis of MTBs currently held at German hospitals showed, that there is less collaborative exploration during the meeting as expected, with a large part of the information search being done during the MTB preparation. Thus we designed our interface to support both situations, a single user preparing the MTB and the presentation of information and group discussion during the meeting. |
Friedrich, Andreas; de la Garza, Luis; Kohlbacher, Oliver Interactive Visualization for Large-Scale Multi-factorial Research Designs Springer, 11371 , 2018. @proceedings{MultiFactorial_DILS2018, title = {Interactive Visualization for Large-Scale Multi-factorial Research Designs}, author = {Andreas Friedrich and Luis de la Garza and Oliver Kohlbacher}, doi = {10.1007/978-3-030-06016-9_7}, year = {2018}, date = {2018-01-01}, volume = {11371}, publisher = {Springer}, series = {Lecture Notes in Computer Science}, keywords = {}, pubstate = {published}, tppubtype = {proceedings} } |
Friedrich, Andreas; de la Garza, Luis; Kohlbacher, Oliver; Nahnsen, Sven Interactive Visualization for Large-Scale Multi-factorial Research Designs Auer, Maria-Esther Vidal Sören (Ed.): DILS 2018: Data Integration in the Life Sciences, pp. 75-84, 2018. @inproceedings{MultFactorial_DILS2018, title = {Interactive Visualization for Large-Scale Multi-factorial Research Designs}, author = {Andreas Friedrich and Luis de la Garza and Oliver Kohlbacher and Sven Nahnsen}, editor = {Maria-Esther Vidal Sören Auer}, url = {https://link.springer.com/chapter/10.1007/978-3-030-06016-9_7}, year = {2018}, date = {2018-01-01}, booktitle = {DILS 2018: Data Integration in the Life Sciences}, volume = {11371}, pages = {75-84}, series = {Lecture Notes in Computer Science}, abstract = {Recent publications have shown that the majority of studies cannot be adequately reproduced. The underlying causes seem to be diverse. Usage of the wrong statistical tools can lead to the reporting of dubious correlations as significant results. Missing information from lab protocols or other metadata can make verification impossible. Especially with the advent of Big Data in the life sciences and the hereby-involved measurement of thousands of multi-omics samples, researchers depend more than ever on adequate metadata annotation. In recent years, the scientific community has created multiple experimental design standards, which try to define the minimum information necessary to make experiments reproducible. Tools help with creation or analysis of this abundance of metadata, but are often still based on spreadsheet formats and lack intuitive visualizations. We present an interactive graph visualization tailored to experiments using a factorial experimental design. Our solution summarizes sample sources and extracted samples based on similarity of independent variables, enabling a quick grasp of the scientific question at the core of the experiment even for large studies. We support the ISA-Tab standard, enabling visualization of diverse omics experiments. As part of our platform for data-driven biomedical research, our implementation offers additional features to detect the status of data generation and more.}, keywords = {}, pubstate = {published}, tppubtype = {inproceedings} } Recent publications have shown that the majority of studies cannot be adequately reproduced. The underlying causes seem to be diverse. Usage of the wrong statistical tools can lead to the reporting of dubious correlations as significant results. Missing information from lab protocols or other metadata can make verification impossible. Especially with the advent of Big Data in the life sciences and the hereby-involved measurement of thousands of multi-omics samples, researchers depend more than ever on adequate metadata annotation. In recent years, the scientific community has created multiple experimental design standards, which try to define the minimum information necessary to make experiments reproducible. Tools help with creation or analysis of this abundance of metadata, but are often still based on spreadsheet formats and lack intuitive visualizations. We present an interactive graph visualization tailored to experiments using a factorial experimental design. Our solution summarizes sample sources and extracted samples based on similarity of independent variables, enabling a quick grasp of the scientific question at the core of the experiment even for large studies. We support the ISA-Tab standard, enabling visualization of diverse omics experiments. As part of our platform for data-driven biomedical research, our implementation offers additional features to detect the status of data generation and more. |
Audain, Enrique; Uszkoreit, Julian; Sachsenberg, Timo; Pfeuffer, Julianus; Liang, Xiao; Hermjakob, Henning; Sanchez, Aniel; Eisenacher, Martin; Reinert, Knut; Tabb, David L; Kohlbacher, Oliver; Perez-Riverol, Yasset In-depth analysis of protein inference algorithms using multiple search engines and well-defined metrics J. Proteomics, 150 , pp. 170–182, 2017. @article{PIA_JProt_2016, title = {In-depth analysis of protein inference algorithms using multiple search engines and well-defined metrics}, author = {Enrique Audain and Julian Uszkoreit and Timo Sachsenberg and Julianus Pfeuffer and Xiao Liang and Henning Hermjakob and Aniel Sanchez and Martin Eisenacher and Knut Reinert and David L Tabb and Oliver Kohlbacher and Yasset Perez-Riverol}, url = {https://doi.org/10.1016/j.jprot.2016.08.002}, year = {2017}, date = {2017-01-01}, journal = {J. Proteomics}, volume = {150}, pages = {170–182}, abstract = {In mass spectrometry-based shotgun proteomics, protein identifications are usually the desired result. However, most of the analytical methods are based on the identification of reliable peptides and not the direct identification of intact proteins. Thus, assembling peptides identified from tandem mass spectra into a list of proteins, referred to as protein inference, is a critical step in proteomics research. Currently, different protein inference algorithms and tools are available for the proteomics community. Here, we evaluated five software tools for protein inference (PIA, ProteinProphet, Fido, ProteinLP, MSBayesPro) using three popular database search engines: Mascot, X!Tandem, and MS-GF+. All the algorithms were evaluated using a highly customizable KNIME workflow using four different public datasets with varying complexities (different sample preparation, species and analytical instruments). We defined a set of quality control metrics to evaluate the performance of each combination of search engines, protein inference algorithm, and parameters on each dataset. We show that the results for complex samples vary not only regarding the actual numbers of reported protein groups but also concerning the actual composition of groups. Furthermore, the robustness of reported proteins when using databases of differing complexities is strongly dependant on the applied inference algorithm. Finally, merging the identifications of multiple search engines does not necessarily increase the number of reported proteins, but does increase the number of peptides per protein and thus can generally be recommended. SIGNIFICANCE: Protein inference is one of the major challenges in MS-based proteomics nowadays. Currently, there are a vast number of protein inference algorithms and implementations available for the proteomics community. Protein assembly impacts in the final results of the research, the quantitation values and the final claims in the research manuscript. Even though protein inference is a crucial step in proteomics data analysis, a comprehensive evaluation of the many different inference methods has never been performed. Previously Journal of proteomics has published multiple studies about other benchmark of bioinformatics algorithms (PMID: 26585461; PMID: 22728601) in proteomics studies making clear the importance of those studies for the proteomics community and the journal audience. This manuscript presents a new bioinformatics solution based on the KNIME/OpenMS platform that aims at providing a fair comparison of protein inference algorithms (https://github.com/KNIME-OMICS). Six different algorithms - ProteinProphet, MSBayesPro, ProteinLP, Fido and PIA- were evaluated using the highly customizable workflow on four public datasets with varying complexities. Five popular database search engines Mascot, X!Tandem, MS-GF+ and combinations thereof were evaluated for every protein inference tool. In total >186 proteins lists were analyzed and carefully compare using three metrics for quality assessments of the protein inference results: 1) the numbers of reported proteins, 2) peptides per protein, and the 3) number of uniquely reported proteins per inference method, to address the quality of each inference method. We also examined how many proteins were reported by choosing each combination of search engines, protein inference algorithms and parameters on each dataset. The results show that using 1) PIA or Fido seems to be a good choice when studying the results of the analyzed workflow, regarding not only the reported proteins and the high-quality identifications, but also the required runtime. 2) Merging the identifications of multiple search engines gives almost always more confident results and increases the number of peptides per protein group. 3) The usage of databases containing not only the canonical, but also known isoforms of proteins has a small impact on the number of reported proteins. The detection of specific isoforms could, concerning the question behind the study, compensate for slightly shorter reports using the parsimonious reports. 4) The current workflow can be easily extended to support new algorithms and search engine combinations.}, keywords = {}, pubstate = {published}, tppubtype = {article} } In mass spectrometry-based shotgun proteomics, protein identifications are usually the desired result. However, most of the analytical methods are based on the identification of reliable peptides and not the direct identification of intact proteins. Thus, assembling peptides identified from tandem mass spectra into a list of proteins, referred to as protein inference, is a critical step in proteomics research. Currently, different protein inference algorithms and tools are available for the proteomics community. Here, we evaluated five software tools for protein inference (PIA, ProteinProphet, Fido, ProteinLP, MSBayesPro) using three popular database search engines: Mascot, X!Tandem, and MS-GF+. All the algorithms were evaluated using a highly customizable KNIME workflow using four different public datasets with varying complexities (different sample preparation, species and analytical instruments). We defined a set of quality control metrics to evaluate the performance of each combination of search engines, protein inference algorithm, and parameters on each dataset. We show that the results for complex samples vary not only regarding the actual numbers of reported protein groups but also concerning the actual composition of groups. Furthermore, the robustness of reported proteins when using databases of differing complexities is strongly dependant on the applied inference algorithm. Finally, merging the identifications of multiple search engines does not necessarily increase the number of reported proteins, but does increase the number of peptides per protein and thus can generally be recommended. SIGNIFICANCE: Protein inference is one of the major challenges in MS-based proteomics nowadays. Currently, there are a vast number of protein inference algorithms and implementations available for the proteomics community. Protein assembly impacts in the final results of the research, the quantitation values and the final claims in the research manuscript. Even though protein inference is a crucial step in proteomics data analysis, a comprehensive evaluation of the many different inference methods has never been performed. Previously Journal of proteomics has published multiple studies about other benchmark of bioinformatics algorithms (PMID: 26585461; PMID: 22728601) in proteomics studies making clear the importance of those studies for the proteomics community and the journal audience. This manuscript presents a new bioinformatics solution based on the KNIME/OpenMS platform that aims at providing a fair comparison of protein inference algorithms (https://github.com/KNIME-OMICS). Six different algorithms - ProteinProphet, MSBayesPro, ProteinLP, Fido and PIA- were evaluated using the highly customizable workflow on four public datasets with varying complexities. Five popular database search engines Mascot, X!Tandem, MS-GF+ and combinations thereof were evaluated for every protein inference tool. In total >186 proteins lists were analyzed and carefully compare using three metrics for quality assessments of the protein inference results: 1) the numbers of reported proteins, 2) peptides per protein, and the 3) number of uniquely reported proteins per inference method, to address the quality of each inference method. We also examined how many proteins were reported by choosing each combination of search engines, protein inference algorithms and parameters on each dataset. The results show that using 1) PIA or Fido seems to be a good choice when studying the results of the analyzed workflow, regarding not only the reported proteins and the high-quality identifications, but also the required runtime. 2) Merging the identifications of multiple search engines gives almost always more confident results and increases the number of peptides per protein group. 3) The usage of databases containing not only the canonical, but also known isoforms of proteins has a small impact on the number of reported proteins. The detection of specific isoforms could, concerning the question behind the study, compensate for slightly shorter reports using the parsimonious reports. 4) The current workflow can be easily extended to support new algorithms and search engine combinations. |
Backert, Linus; Kowalewski, Daniel; Walz, Simon; Schuster, Heiko; Berlin, Claudia; Neidert, Marian; Schemionek, Mirle; Brümmendorf, Tim Hendrik; Vicinic, Vladan; Niederwieser, Dietger; Kanz, Lothar; Salih, Helmut Rainer; Kohlbacher, Oliver; Weisel, Katja; Rammensee, Hans-Georg; Stevanovic, Stefan; Walz, Juliana Sarah A meta-analysis of HLA peptidome composition in different hematological entities: Entity-specific dividing lines and Oncotarget, 8 (27), pp. 43915-43924., 2017. @article{panLeukemiaAntigens_Oncotargets2016, title = {A meta-analysis of HLA peptidome composition in different hematological entities: Entity-specific dividing lines and }, author = {Linus Backert and Daniel Kowalewski and Simon Walz and Heiko Schuster and Claudia Berlin and Marian Neidert and Mirle Schemionek and Tim Hendrik Brümmendorf and Vladan Vicinic and Dietger Niederwieser and Lothar Kanz and Helmut Rainer Salih and Oliver Kohlbacher and Katja Weisel and Hans-Georg Rammensee and Stefan Stevanovic and Juliana Sarah Walz}, url = {https://doi.org/10.18632/oncotarget.14918}, year = {2017}, date = {2017-01-01}, journal = {Oncotarget}, volume = {8}, number = {27}, pages = {43915-43924.}, abstract = {Hematological malignancies (HM) are highly amenable targets for immunotherapeutic intervention and may be effectively treated by antigen-specific T-cell based treatment. Recent studies demonstrate that physiologically occurring anti-cancer T-cell responses in certain HM entities target broadly presented non-mutated epitopes. HLA ligands are thus implied as prime targets for broadly applicable and antigen-specific off-the-shelf compounds. With the aim of assessing the presence of common targets shared among different HM which may enable addressing a larger patient collective we conducted a meta-analysis of 83 mass spectrometry-based HLA peptidome datasets (comprising 40,361 unique peptide identifications) across four major HM (19 AML, 16 CML, 35 CLL, and 13 MM/MCL samples) and investigated similarities and differences within the HLA presented antigenic landscape. We found the cancer HLA peptidome datasets to cluster specifically along entity and lineage lines, suggesting that the immunopeptidome directly reflects the differences in the underlying (tumor-)biology. In line with these findings, we only detected a small set of entity-spanning antigens, which were predominantly characterized by low presentation frequencies within the different patient cohorts. These findings suggest that design of T-cell immunotherapies for the treatment of HM should ideally be conducted in an entity-specific fashion.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Hematological malignancies (HM) are highly amenable targets for immunotherapeutic intervention and may be effectively treated by antigen-specific T-cell based treatment. Recent studies demonstrate that physiologically occurring anti-cancer T-cell responses in certain HM entities target broadly presented non-mutated epitopes. HLA ligands are thus implied as prime targets for broadly applicable and antigen-specific off-the-shelf compounds. With the aim of assessing the presence of common targets shared among different HM which may enable addressing a larger patient collective we conducted a meta-analysis of 83 mass spectrometry-based HLA peptidome datasets (comprising 40,361 unique peptide identifications) across four major HM (19 AML, 16 CML, 35 CLL, and 13 MM/MCL samples) and investigated similarities and differences within the HLA presented antigenic landscape. We found the cancer HLA peptidome datasets to cluster specifically along entity and lineage lines, suggesting that the immunopeptidome directly reflects the differences in the underlying (tumor-)biology. In line with these findings, we only detected a small set of entity-spanning antigens, which were predominantly characterized by low presentation frequencies within the different patient cohorts. These findings suggest that design of T-cell immunotherapies for the treatment of HM should ideally be conducted in an entity-specific fashion. |
Haen, Sebastian P; Groh, Christiane; Schumm, Michael; Backert, Linus; Löffler, Markus W; Federmann, Birgit; Faul, Christoph; Dörfel, Daniela; Vogel, Wichard; Handgretinger, Rupert; Kanz, Lothar; Bethge, Wolfgang A Haploidentical hematopoietic cell transplantation using in vitro T cell depleted grafts as salvage therapy in patients with disease relapse after prior allogeneic transplantation Annals of Hematology, pp. 1-11, 2017. @article{Haen2017, title = {Haploidentical hematopoietic cell transplantation using in vitro T cell depleted grafts as salvage therapy in patients with disease relapse after prior allogeneic transplantation}, author = {Sebastian P Haen and Christiane Groh and Michael Schumm and Linus Backert and Markus W Löffler and Birgit Federmann and Christoph Faul and Daniela Dörfel and Wichard Vogel and Rupert Handgretinger and Lothar Kanz and Wolfgang A Bethge}, url = {http://dx.doi.org/10.1007/s00277-017-2941-x}, year = {2017}, date = {2017-01-01}, journal = {Annals of Hematology}, pages = {1-11}, abstract = {Disease relapse after one or more allogeneic hematopoietic cell transplantations (HCT) represents a therapeutic challenge with all options bearing a significant morbidity and mortality. Haploidentical HCT may induce more pronounced anti-leukemic effects and was evaluated at our center in 25 consecutive patients with disease relapse after preceding HCT receiving haploidentical grafts after in vitro T cell depletion. Overall survival at 1 and 2 years was 32 and 14%, respectively. Of note, patients with complete remission (CR) before haploidentical HCT had a very favorable overall survival of 41.7% at 2 years. Cumulative incidence of non-relapse mortality was 36 and 40% at 1 and 2 years, respectively. With a cumulative incidence for relapse of 36 and 45.6% at 1 and 2 years, disease-free survival (DFS) was 28 and 14.4%, respectively. Here also, patients with CR before haploidentical HCT had a favorable DFS of 42% at 2 years. Only very limited acute (11 patients (44%) with a median grade 1) and chronic graft versus host disease (GvHD) (5 patients (11%), limited grade only) was observed. The main complications and causes of death comprised - besides relapse - infections and bleeding complications. Hence, haploidentical HCT can achieve long-term survival comparable to second transplantation with matched or mismatched donors for patients with otherwise deleterious prognosis and should be considered as a treatment option for patients experiencing disease relapse after previous allogeneic HCT.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Disease relapse after one or more allogeneic hematopoietic cell transplantations (HCT) represents a therapeutic challenge with all options bearing a significant morbidity and mortality. Haploidentical HCT may induce more pronounced anti-leukemic effects and was evaluated at our center in 25 consecutive patients with disease relapse after preceding HCT receiving haploidentical grafts after in vitro T cell depletion. Overall survival at 1 and 2 years was 32 and 14%, respectively. Of note, patients with complete remission (CR) before haploidentical HCT had a very favorable overall survival of 41.7% at 2 years. Cumulative incidence of non-relapse mortality was 36 and 40% at 1 and 2 years, respectively. With a cumulative incidence for relapse of 36 and 45.6% at 1 and 2 years, disease-free survival (DFS) was 28 and 14.4%, respectively. Here also, patients with CR before haploidentical HCT had a favorable DFS of 42% at 2 years. Only very limited acute (11 patients (44%) with a median grade 1) and chronic graft versus host disease (GvHD) (5 patients (11%), limited grade only) was observed. The main complications and causes of death comprised - besides relapse - infections and bleeding complications. Hence, haploidentical HCT can achieve long-term survival comparable to second transplantation with matched or mismatched donors for patients with otherwise deleterious prognosis and should be considered as a treatment option for patients experiencing disease relapse after previous allogeneic HCT. |
Nelde, Annika; Kowalewski, Daniel J; Backert, Linus; Schuster, Heiko; Werner, Jan-Ole; Klein, Reinhild; Kohlbacher, Oliver; Kanz, Lothar; Salih, Helmut R; Rammensee, Hans-Georg; Stevanović, Stefan; Stickel, Juliane S HLA ligandome analysis of primary chronic lymphocytic leukemia (CLL) cells under lenalidomide treatment confirms the suitability of lenalidomide for combination with T-cell based immunotherapy OncoImmunol., 5 (12), pp. e1249560, 2017. @article{Walz_OncoImmuno_2017, title = {HLA ligandome analysis of primary chronic lymphocytic leukemia (CLL) cells under lenalidomide treatment confirms the suitability of lenalidomide for combination with T-cell based immunotherapy}, author = {Annika Nelde and Daniel J Kowalewski and Linus Backert and Heiko Schuster and Jan-Ole Werner and Reinhild Klein and Oliver Kohlbacher and Lothar Kanz and Helmut R Salih and Hans-Georg Rammensee and Stefan Stevanović and Juliane S Stickel}, url = {https://doi.org/10.1080/2162402X.2017.1316438}, year = {2017}, date = {2017-01-01}, journal = {OncoImmunol.}, volume = {5}, number = {12}, pages = {e1249560}, abstract = {We recently completed a phase I/IIa trial of RNActive® CV9201, a novel mRNA-based therapeutic vaccine targeting five tumor-associated antigens in non-small cell lung cancer (NSCLC) patients. The aim of the study presented here was to comprehensively analyze changes in peripheral blood during the vaccination period and to generate hypotheses facilitating the identification of potential biomarkers correlating with differential clinical outcomes post RNActive® immunotherapy. We performed whole-genome expression profiling in a subgroup of 22 stage IV NSCLC patients before and after initiation of treatment with CV9201. Utilizing an analytic approach based on blood transcriptional modules (BTMs), a previously described, sensitive tool for blood transcriptome data analysis, patients segregated into two major clusters based on transcriptional changes post RNActive® treatment. The first group of patients was characterized by the upregulation of an expression signature associated with myeloid cells and inflammation, whereas the other group exhibited an expression signature associated with T and NK cells. Patients with an enrichment of T and NK cell modules after treatment compared to baseline exhibited significantly longer progression-free and overall survival compared to patients with an upregulation of myeloid cell and inflammatory modules. Notably, these gene expression signatures were mutually exclusive and inversely correlated. Furthermore, our findings correlated with phenotypic data derived by flow cytometry as well as the neutrophil-to-lymphocyte ratio. Our study thus demonstrates non-overlapping, distinct transcriptional profiles correlating with survival warranting further validation for the development of biomarker candidates for mRNA-based immunotherapy.}, keywords = {}, pubstate = {published}, tppubtype = {article} } We recently completed a phase I/IIa trial of RNActive® CV9201, a novel mRNA-based therapeutic vaccine targeting five tumor-associated antigens in non-small cell lung cancer (NSCLC) patients. The aim of the study presented here was to comprehensively analyze changes in peripheral blood during the vaccination period and to generate hypotheses facilitating the identification of potential biomarkers correlating with differential clinical outcomes post RNActive® immunotherapy. We performed whole-genome expression profiling in a subgroup of 22 stage IV NSCLC patients before and after initiation of treatment with CV9201. Utilizing an analytic approach based on blood transcriptional modules (BTMs), a previously described, sensitive tool for blood transcriptome data analysis, patients segregated into two major clusters based on transcriptional changes post RNActive® treatment. The first group of patients was characterized by the upregulation of an expression signature associated with myeloid cells and inflammation, whereas the other group exhibited an expression signature associated with T and NK cells. Patients with an enrichment of T and NK cell modules after treatment compared to baseline exhibited significantly longer progression-free and overall survival compared to patients with an upregulation of myeloid cell and inflammatory modules. Notably, these gene expression signatures were mutually exclusive and inversely correlated. Furthermore, our findings correlated with phenotypic data derived by flow cytometry as well as the neutrophil-to-lymphocyte ratio. Our study thus demonstrates non-overlapping, distinct transcriptional profiles correlating with survival warranting further validation for the development of biomarker candidates for mRNA-based immunotherapy. |
Heimgärtner, Florian; Hettich, Stefan; Kohlbacher, Oliver; Menth, Michael Scaling Home Automation to Public Buildings: A Distributed Multiuser Setup for OpenHAB 2 Global Internet of Things Summit (GIoTS) 2017, 2017. @conference{GIoTS2017, title = {Scaling Home Automation to Public Buildings: A Distributed Multiuser Setup for OpenHAB 2}, author = {Florian Heimgärtner and Stefan Hettich and Oliver Kohlbacher and Michael Menth}, url = {http://dx.doi.org/10.1109/GIOTS.2017.8016235}, year = {2017}, date = {2017-01-01}, booktitle = {Global Internet of Things Summit (GIoTS) 2017}, abstract = {Home automation systems can help to reduce energy costs and increase comfort of living by adjusting room temperatures according to schedules, rules, and sensor input. OpenHAB 2 is an open-source home automation framework supporting various home automation technologies and devices. While OpenHAB is well suited for single occupancy homes, large public buildings pose additional challenges. The limited range of wireless home automation technologies requires transceivers distributed across the building. Additionally, control permissions need to be restricted to authorized persons. This work presents OpenHAB-DM, a distributed OpenHAB 2 setup with extensions introducing user authentication, access control, and management tools for decentralized OpenHAB node deployment.}, keywords = {}, pubstate = {published}, tppubtype = {conference} } Home automation systems can help to reduce energy costs and increase comfort of living by adjusting room temperatures according to schedules, rules, and sensor input. OpenHAB 2 is an open-source home automation framework supporting various home automation technologies and devices. While OpenHAB is well suited for single occupancy homes, large public buildings pose additional challenges. The limited range of wireless home automation technologies requires transceivers distributed across the building. Additionally, control permissions need to be restricted to authorized persons. This work presents OpenHAB-DM, a distributed OpenHAB 2 setup with extensions introducing user authentication, access control, and management tools for decentralized OpenHAB node deployment. |
Schubert, Benjamin; de la Garza, Luis; Mohr, Christopher; Walzer, Mathias; Kohlbacher, Oliver ImmunoNodes - Graphical Development of Complex Immunoinformatics Workflows BMC Bioinformatics, 18 (1), pp. 242, 2017. @article{ImmunoNodes2017, title = {ImmunoNodes - Graphical Development of Complex Immunoinformatics Workflows}, author = {Benjamin Schubert and Luis de la Garza and Christopher Mohr and Mathias Walzer and Oliver Kohlbacher}, url = {https://doi.org/10.1186/s12859-017-1667-z}, year = {2017}, date = {2017-01-01}, journal = {BMC Bioinformatics}, volume = {18}, number = {1}, pages = {242}, abstract = {BACKGROUND: Immunoinformatics has become a crucial part in biomedical research. Yet many immunoinformatics tools have command line interfaces only and can be difficult to install. Web-based immunoinformatics tools, on the other hand, are difficult to integrate with other tools, which is typically required for the complex analysis and prediction pipelines required for advanced applications. RESULT: We present ImmunoNodes, an immunoinformatics toolbox that is fully integrated into the visual workflow environment KNIME. By dragging and dropping tools and connecting them to indicate the data flow through the pipeline, it is possible to construct very complex workflows without the need for coding. CONCLUSION: ImmunoNodes allows users to build complex workflows with an easy to use and intuitive interface with a few clicks on any desktop computer.}, keywords = {}, pubstate = {published}, tppubtype = {article} } BACKGROUND: Immunoinformatics has become a crucial part in biomedical research. Yet many immunoinformatics tools have command line interfaces only and can be difficult to install. Web-based immunoinformatics tools, on the other hand, are difficult to integrate with other tools, which is typically required for the complex analysis and prediction pipelines required for advanced applications. RESULT: We present ImmunoNodes, an immunoinformatics toolbox that is fully integrated into the visual workflow environment KNIME. By dragging and dropping tools and connecting them to indicate the data flow through the pipeline, it is possible to construct very complex workflows without the need for coding. CONCLUSION: ImmunoNodes allows users to build complex workflows with an easy to use and intuitive interface with a few clicks on any desktop computer. |
Vizcaino, Juan Antonio; Mayer, Gerhard; Perkins, Simon R; Barnsnes, Harald; Vaudel, Marc; Perez-Riverol, Yasset; Ternent, Tobias; Uszkoreit, Julian; Eisenacher, Martin; Fischer, Lutz; Rappsilber, Juri; Netz, Eugen; Walzer, Mathias; Kohlbacher, Oliver; Leitner, Alexander; Chalkley, Robert J; Ghali, Fawaz; Martínez-Bartolomé, Salvador; Deutsch, Eric W; Jones, Andrew R The mzIdentML data standard version 1.2, supporting advances in proteome informatics Mol. Cell. Prot., 16 (7), pp. 1275-1285, 2017. @article{mzIdML1.2_2017, title = {The mzIdentML data standard version 1.2, supporting advances in proteome informatics}, author = {Juan Antonio Vizcaino and Gerhard Mayer and Simon R Perkins and Harald Barnsnes and Marc Vaudel and Yasset Perez-Riverol and Tobias Ternent and Julian Uszkoreit and Martin Eisenacher and Lutz Fischer and Juri Rappsilber and Eugen Netz and Mathias Walzer and Oliver Kohlbacher and Alexander Leitner and Robert J Chalkley and Fawaz Ghali and Salvador Martínez-Bartolomé and Eric W Deutsch and Andrew R Jones}, url = {https://doi.org/10.1074/mcp.M117.068429}, year = {2017}, date = {2017-01-01}, journal = {Mol. Cell. Prot.}, volume = {16}, number = {7}, pages = {1275-1285}, abstract = {The first stable version of the Proteomics Standards Initiative mzIdentML open data standard (version 1.1) was published in 2012 - capturing the outputs of peptide and protein identification software. In the intervening years, the standard has become well supported in both commercial and open software, as well as a submission and download format for public repositories. Here we report a new release of mzIdentML (version 1.2) that is required to keep pace with emerging practice in proteome informatics. New features have been added to support: (i) scores associated with localization of modifications on peptides; (ii) statistics performed at the level of peptides; (iii) identification of cross-linked peptides; and (iv) support for proteogenomics approaches. In addition, there is now improved support for the encoding of de novo sequencing of peptides, spectral library searches and protein inference. As a key point, the underlying XML schema has only undergone very minor modifications to simplify as much as possible the transition from version 1.1 to version 1.2 for implementers, but there have been several notable updates to the format specification, implementation guidelines, controlled vocabularies and validation software. mzIdentML 1.2 can be described as backwards compatible, in that reading software designed for mzIdentML 1.1 should function in most cases without adaptation. We anticipate that these developments will provide a continued stable base for software teams working to implement the standard. All the related documentation is accessible at http://www.psidev.info/mzidentml.}, keywords = {}, pubstate = {published}, tppubtype = {article} } The first stable version of the Proteomics Standards Initiative mzIdentML open data standard (version 1.1) was published in 2012 - capturing the outputs of peptide and protein identification software. In the intervening years, the standard has become well supported in both commercial and open software, as well as a submission and download format for public repositories. Here we report a new release of mzIdentML (version 1.2) that is required to keep pace with emerging practice in proteome informatics. New features have been added to support: (i) scores associated with localization of modifications on peptides; (ii) statistics performed at the level of peptides; (iii) identification of cross-linked peptides; and (iv) support for proteogenomics approaches. In addition, there is now improved support for the encoding of de novo sequencing of peptides, spectral library searches and protein inference. As a key point, the underlying XML schema has only undergone very minor modifications to simplify as much as possible the transition from version 1.1 to version 1.2 for implementers, but there have been several notable updates to the format specification, implementation guidelines, controlled vocabularies and validation software. mzIdentML 1.2 can be described as backwards compatible, in that reading software designed for mzIdentML 1.1 should function in most cases without adaptation. We anticipate that these developments will provide a continued stable base for software teams working to implement the standard. All the related documentation is accessible at http://www.psidev.info/mzidentml. |
Pfeuffer, Julianus U; Sachsenberg, Timo; Alka, Oliver; Walzer, Mathias; Fillbrunn, Alexander; Nilse, Lars; Schilling, Oliver; Reinert, Knut; Kohlbacher, Oliver OpenMS - A platform for reproducible analysis of mass spectrometry data J. Biotechnol., pp. S0168-1656(17)30251-1, 2017. @article{OpenMSJBiotech2017, title = {OpenMS - A platform for reproducible analysis of mass spectrometry data}, author = {Julianus U Pfeuffer and Timo Sachsenberg and Oliver Alka and Mathias Walzer and Alexander Fillbrunn and Lars Nilse and Oliver Schilling and Knut Reinert and Oliver Kohlbacher}, url = {https://doi.org/10.1016/j.jbiotec.2017.05.016}, year = {2017}, date = {2017-01-01}, journal = {J. Biotechnol.}, pages = {S0168-1656(17)30251-1}, abstract = {In recent years, several mass spectrometry-based omics technologies emerged to investigate qualitative and quantitative changes within thousands of biologically active components such as proteins, lipids and metabolites. The research enabled through these methods potentially contributes to the diagnosis and pathophysiology of human diseases as well as to the clarification of structures and interactions between biomolecules. Simultaneously, technological advances in the field of mass spectrometry leading to an ever increasing amount of data, demand high standards in efficiency, accuracy and reproducibility of potential analysis software. RESULTS: This article presents the current state and ongoing developments in OpenMS, a versatile open-source framework aimed at enabling reproducible analyses of high-throughput mass spectrometry data. It provides implementations of frequently occurring processing operations on MS data through a clean application programming interface in C++ and Python. A collection of 185 tools and ready-made workflows for typical MS-based experiments enable convenient analyses for non-developers and facilitate reproducible research without losing flexibility. CONCLUSIONS: OpenMS will continue to increase its ease of use for developers as well as users with improved continuous integration/deployment strategies, regular trainings with updated training materials and multiple sources of support. The active developer community ensures the incorporation of new features to support state of the art research.}, keywords = {}, pubstate = {published}, tppubtype = {article} } In recent years, several mass spectrometry-based omics technologies emerged to investigate qualitative and quantitative changes within thousands of biologically active components such as proteins, lipids and metabolites. The research enabled through these methods potentially contributes to the diagnosis and pathophysiology of human diseases as well as to the clarification of structures and interactions between biomolecules. Simultaneously, technological advances in the field of mass spectrometry leading to an ever increasing amount of data, demand high standards in efficiency, accuracy and reproducibility of potential analysis software. RESULTS: This article presents the current state and ongoing developments in OpenMS, a versatile open-source framework aimed at enabling reproducible analyses of high-throughput mass spectrometry data. It provides implementations of frequently occurring processing operations on MS data through a clean application programming interface in C++ and Python. A collection of 185 tools and ready-made workflows for typical MS-based experiments enable convenient analyses for non-developers and facilitate reproducible research without losing flexibility. CONCLUSIONS: OpenMS will continue to increase its ease of use for developers as well as users with improved continuous integration/deployment strategies, regular trainings with updated training materials and multiple sources of support. The active developer community ensures the incorporation of new features to support state of the art research. |
Chevrette, Marc G; Aicheler, Fabian; Kohlbacher, Oliver; Currie, Cameron R; Medema, Marnix H SANDPUMA: Ensemble Predictions of Nonribosomal Peptide Chemistry Reveals Biosynthetic Diversity across Actinobacteria Bioinformatics, 33 (20), pp. 3202-3210, 2017. @article{SANDPUMA2017, title = {SANDPUMA: Ensemble Predictions of Nonribosomal Peptide Chemistry Reveals Biosynthetic Diversity across Actinobacteria}, author = {Marc G Chevrette and Fabian Aicheler and Oliver Kohlbacher and Cameron R Currie and Marnix H Medema}, url = {https://doi.org/10.1093/bioinformatics/btx400}, year = {2017}, date = {2017-01-01}, journal = {Bioinformatics}, volume = {33}, number = {20}, pages = {3202-3210}, abstract = {Summary: Nonribosomally synthesized peptides (NRPs) are natural products with widespread applications in medicine and biotechnology. Many algorithms have been developed to predict the substrate specificities of nonribosomal peptide synthetase adenylation (A) domains from DNA sequences, which enables prioritization and dereplication, and integration with other data types in discovery efforts. However, insufficient training data and a lack of clarity regarding prediction quality have impeded optimal use. Here, we introduce prediCAT, a new phylogenetics-inspired algorithm, which quantitatively estimates the degree of predictability of each A-domain. We then systematically benchmarked all algorithms on a newly-gathered, independent test set of 434 A-domain sequences, showing that active-site-motif-based algorithms outperform whole-domain-based methods. Subsequently, we developed SANDPUMA, a powerful ensemble algorithm, based on newly-trained versions of all high-performing algorithms, which significantly outperforms individual methods. Finally, we deployed SANDPUMA in a systematic investigation of 7,635 Actinobacteria genomes, suggesting that NRP chemical diversity is much higher than previously estimated. SANDPUMA has been integrated into the widely-used antiSMASH biosynthetic gene cluster analysis pipeline and is also available as an open-source, standalone tool. Availability: SANDPUMA is freely available at https://bitbucket.org/chevrm/sandpuma and as a docker image at https://hub.docker.com/r/chevrm/sandpuma/ under the GNU Public License 3 (GPL3). Contact: , }, keywords = {}, pubstate = {published}, tppubtype = {article} } Summary: Nonribosomally synthesized peptides (NRPs) are natural products with widespread applications in medicine and biotechnology. Many algorithms have been developed to predict the substrate specificities of nonribosomal peptide synthetase adenylation (A) domains from DNA sequences, which enables prioritization and dereplication, and integration with other data types in discovery efforts. However, insufficient training data and a lack of clarity regarding prediction quality have impeded optimal use. Here, we introduce prediCAT, a new phylogenetics-inspired algorithm, which quantitatively estimates the degree of predictability of each A-domain. We then systematically benchmarked all algorithms on a newly-gathered, independent test set of 434 A-domain sequences, showing that active-site-motif-based algorithms outperform whole-domain-based methods. Subsequently, we developed SANDPUMA, a powerful ensemble algorithm, based on newly-trained versions of all high-performing algorithms, which significantly outperforms individual methods. Finally, we deployed SANDPUMA in a systematic investigation of 7,635 Actinobacteria genomes, suggesting that NRP chemical diversity is much higher than previously estimated. SANDPUMA has been integrated into the widely-used antiSMASH biosynthetic gene cluster analysis pipeline and is also available as an open-source, standalone tool. Availability: SANDPUMA is freely available at https://bitbucket.org/chevrm/sandpuma and as a docker image at https://hub.docker.com/r/chevrm/sandpuma/ under the GNU Public License 3 (GPL3). Contact: , |
Vizcaíno, Juan Antonio; Walzer, Mathias; Jiménez, Rafael C; Bittremieux, Wout; Bouyssie, David; Carapito, Christine; Corrales, Fernando; Ferro, Myriam; Heck, Albert J R; Horvatovich, Peter; Hubalek, Martin; Lane, Lydia; Laukens, Kris; Levander, Fredrik; Lisacek, Frederique; Novak, Petr; Palmblad, Magnus; Piovesan, Damiano; Pühler, Alfred; Schwämmle, Veit; Valkenborg, Dirk; van Rijswijk, Merlijn; Vondrasek, Jiri; Eisenacher, Martin; Martens, Lennart; Kohlbacher, Oliver A community proposal to integrate proteomics activities in ELIXIR F1000Research, 6 , pp. 875, 2017. @article{ProteomicsELIXIR2017, title = {A community proposal to integrate proteomics activities in ELIXIR}, author = {Juan Antonio Vizcaíno and Mathias Walzer and Rafael C Jiménez and Wout Bittremieux and David Bouyssie and Christine Carapito and Fernando Corrales and Myriam Ferro and Albert J R Heck and Peter Horvatovich and Martin Hubalek and Lydia Lane and Kris Laukens and Fredrik Levander and Frederique Lisacek and Petr Novak and Magnus Palmblad and Damiano Piovesan and Alfred Pühler and Veit Schwämmle and Dirk Valkenborg and Merlijn van Rijswijk and Jiri Vondrasek and Martin Eisenacher and Lennart Martens and Oliver Kohlbacher}, url = {https://doi.org/10.12688/f1000research.11751.1}, year = {2017}, date = {2017-01-01}, journal = {F1000Research}, volume = {6}, pages = {875}, abstract = {Computational approaches have been major drivers behind the progress of proteomics in recent years. The aim of this white paper is to provide a framework for integrating computational proteomics into ELIXIR in the near future, and thus to broaden the portfolio of omics technologies supported by this European distributed infrastructure. This white paper is the direct result of a strategy meeting on ‘The Future of Proteomics in ELIXIR’ that took place in March 2017 in Tübingen (Germany), and involved representatives of eleven ELIXIR nodes. These discussions led to a list of priority areas in computational proteomics that would complement existing activities and close gaps in the portfolio of tools and services offered by ELIXIR so far. We provide some suggestions on how these activities could be integrated into ELIXIR’s existing platforms, and how it could lead to a new ELIXIR use case in proteomics. We also highlight connections to the related field of metabolomics, where similar activities are ongoing. This white paper could thus serve as a starting point for the integration of computational proteomics into ELIXIR. Over the next few months we will be working closely with all stakeholders involved, and in particular with other representatives of the proteomics community, to further refine this paper.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Computational approaches have been major drivers behind the progress of proteomics in recent years. The aim of this white paper is to provide a framework for integrating computational proteomics into ELIXIR in the near future, and thus to broaden the portfolio of omics technologies supported by this European distributed infrastructure. This white paper is the direct result of a strategy meeting on ‘The Future of Proteomics in ELIXIR’ that took place in March 2017 in Tübingen (Germany), and involved representatives of eleven ELIXIR nodes. These discussions led to a list of priority areas in computational proteomics that would complement existing activities and close gaps in the portfolio of tools and services offered by ELIXIR so far. We provide some suggestions on how these activities could be integrated into ELIXIR’s existing platforms, and how it could lead to a new ELIXIR use case in proteomics. We also highlight connections to the related field of metabolomics, where similar activities are ongoing. This white paper could thus serve as a starting point for the integration of computational proteomics into ELIXIR. Over the next few months we will be working closely with all stakeholders involved, and in particular with other representatives of the proteomics community, to further refine this paper. |
Alberer, Martin; Gnad-Vogt, Ulrike; Hong, Henoch Sangjoon; Mehr, Keyvan Tadjalli; Backert, Linus; Finak, Greg; Gottardo, Raphael; Bica, Mihai Alexandru; Garofano, Aurelio; Koch, Sven Dominik; Fotin-Mleczek, Mariola; Hoerr, Ingmar; Clemens, Ralf; von Sonnenburg, Frank Safety and immunogenicity of a mRNA rabies vaccine in healthy adults: an open-label, non-randomised, prospective, first-in-human phase 1 clinical trial The Lancet, 390 (10101), pp. 1511-1520, 2017. @article{Alberer2017, title = {Safety and immunogenicity of a mRNA rabies vaccine in healthy adults: an open-label, non-randomised, prospective, first-in-human phase 1 clinical trial}, author = {Martin Alberer and Ulrike Gnad-Vogt and Henoch Sangjoon Hong and Keyvan Tadjalli Mehr and Linus Backert and Greg Finak and Raphael Gottardo and Mihai Alexandru Bica and Aurelio Garofano and Sven Dominik Koch and Mariola Fotin-Mleczek and Ingmar Hoerr and Ralf Clemens and Frank von Sonnenburg}, url = {http://linkinghub.elsevier.com/retrieve/pii/S0140673617316653}, year = {2017}, date = {2017-01-01}, journal = {The Lancet}, volume = {390}, number = {10101}, pages = {1511-1520}, abstract = {BACKGROUND: Vaccines based on mRNA coding for antigens have been shown to be safe and immunogenic in preclinical models. We aimed to report results of the first-in-human proof-of-concept clinical trial in healthy adults of a prophylactic mRNA-based vaccine encoding rabies virus glycoprotein (CV7201). METHODS: We did an open-label, uncontrolled, prospective, phase 1 clinical trial at one centre in Munich, Germany. Healthy male and female volunteers (aged 18-40 years) with no history of rabies vaccination were sequentially enrolled. They received three doses of CV7201 intradermally or intramuscularly by needle-syringe or one of three needle-free devices. Escalating doses were given to subsequent cohorts, and one cohort received a booster dose after 1 year. The primary endpoint was safety and tolerability. The secondary endpoint was to determine the lowest dose of CV7201 to elicit rabies virus neutralising titres equal to or greater than the WHO-specified protective antibody titre of 0·5 IU/mL. The study is continuing for long-term safety and immunogenicity follow-up. This trial is registered with ClinicalTrials.gov, number NCT02241135. FINDINGS: Between Oct 21, 2013, and Jan 11, 2016, we enrolled and vaccinated 101 participants with 306 doses of mRNA (80-640 μg) by needle-syringe (18 intradermally and 24 intramuscularly) or needle-free devices (46 intradermally and 13 intramuscularly). In the 7 days post vaccination, 60 (94%) of 64 intradermally vaccinated participants and 36 (97%) of 37 intramuscularly vaccinated participants reported solicited injection site reactions, and 50 (78%) of 64 intradermally vaccinated participants and 29 (78%) of 37 intramuscularly vaccinated participants reported solicited systemic adverse events, including ten grade 3 events. One unexpected, possibly related, serious adverse reaction that occurred 7 days after a 640 μg intramuscular dose resolved without sequelae. mRNA vaccination by needle-free intradermal or intramuscular device injection induced virus neutralising antibody titres of 0·5 IU/mL or more across dose levels and schedules in 32 (71%) of 45 participants given 80 μg or 160 μg CV7201 doses intradermally and six (46%) of 13 participants given 200 μg or 400 μg CV7201 doses intramuscularly. 1 year later, eight (57%) of 14 participants boosted with an 80 μg needle-free intradermal dose of CV7201 achieved titres of 0·5 IU/mL or more. Conversely, intradermal or intramuscular needle-syringe injection was ineffective, with only one participant (who received 320 μg intradermally) showing a detectable immune response. INTERPRETATION: This first-ever demonstration in human beings shows that a prophylactic mRNA-based candidate vaccine can induce boostable functional antibodies against a viral antigen when administered with a needle-free device, although not when injected by a needle-syringe. The vaccine was generally safe with a reasonable tolerability profile. FUNDING: CureVac AG.}, keywords = {}, pubstate = {published}, tppubtype = {article} } BACKGROUND: Vaccines based on mRNA coding for antigens have been shown to be safe and immunogenic in preclinical models. We aimed to report results of the first-in-human proof-of-concept clinical trial in healthy adults of a prophylactic mRNA-based vaccine encoding rabies virus glycoprotein (CV7201). METHODS: We did an open-label, uncontrolled, prospective, phase 1 clinical trial at one centre in Munich, Germany. Healthy male and female volunteers (aged 18-40 years) with no history of rabies vaccination were sequentially enrolled. They received three doses of CV7201 intradermally or intramuscularly by needle-syringe or one of three needle-free devices. Escalating doses were given to subsequent cohorts, and one cohort received a booster dose after 1 year. The primary endpoint was safety and tolerability. The secondary endpoint was to determine the lowest dose of CV7201 to elicit rabies virus neutralising titres equal to or greater than the WHO-specified protective antibody titre of 0·5 IU/mL. The study is continuing for long-term safety and immunogenicity follow-up. This trial is registered with ClinicalTrials.gov, number NCT02241135. FINDINGS: Between Oct 21, 2013, and Jan 11, 2016, we enrolled and vaccinated 101 participants with 306 doses of mRNA (80-640 μg) by needle-syringe (18 intradermally and 24 intramuscularly) or needle-free devices (46 intradermally and 13 intramuscularly). In the 7 days post vaccination, 60 (94%) of 64 intradermally vaccinated participants and 36 (97%) of 37 intramuscularly vaccinated participants reported solicited injection site reactions, and 50 (78%) of 64 intradermally vaccinated participants and 29 (78%) of 37 intramuscularly vaccinated participants reported solicited systemic adverse events, including ten grade 3 events. One unexpected, possibly related, serious adverse reaction that occurred 7 days after a 640 μg intramuscular dose resolved without sequelae. mRNA vaccination by needle-free intradermal or intramuscular device injection induced virus neutralising antibody titres of 0·5 IU/mL or more across dose levels and schedules in 32 (71%) of 45 participants given 80 μg or 160 μg CV7201 doses intradermally and six (46%) of 13 participants given 200 μg or 400 μg CV7201 doses intramuscularly. 1 year later, eight (57%) of 14 participants boosted with an 80 μg needle-free intradermal dose of CV7201 achieved titres of 0·5 IU/mL or more. Conversely, intradermal or intramuscular needle-syringe injection was ineffective, with only one participant (who received 320 μg intradermally) showing a detectable immune response. INTERPRETATION: This first-ever demonstration in human beings shows that a prophylactic mRNA-based candidate vaccine can induce boostable functional antibodies against a viral antigen when administered with a needle-free device, although not when injected by a needle-syringe. The vaccine was generally safe with a reasonable tolerability profile. FUNDING: CureVac AG. |
van Rijswijk, M; Beirnaert, C; Caron, C; Cascante, M; Dominguez, V; Dunn, WB; Ebbels, TMD; Giacomoni, F; Gonzalez-Beltran, A; Hankemeier, T; Haug, K; Izquierdo-Garcia, JL; Jimenez, RC; Jourdan, F; Kale, N; Klapa, MI; Kohlbacher, O; Koort, K; Kultima, K; Corguillé, Le G; Moschonas, NK; Neumann, S; O?Donovan, C; Reczko, M; Rocca-Serra, P; Rosato, A; Salek, RM; Sansone, SA; Satagopam, V; Schober, D; Shimmo, R; Spicer, RA; Spjuth, O; Thévenot, EA; Viant, MR; Weber, RJM; Willighagen, EL; Zanetti, G; Steinbeck, C The future of metabolomics in ELIXIR [version 1; referees: awaiting peer review] F1000Research, 6 (1649), 2017. @article{10.12688-f1000research.12342.2, title = {The future of metabolomics in ELIXIR [version 1; referees: awaiting peer review]}, author = {M van Rijswijk and C Beirnaert and C Caron and M Cascante and V Dominguez and WB Dunn and TMD Ebbels and F Giacomoni and A Gonzalez-Beltran and T Hankemeier and K Haug and JL Izquierdo-Garcia and RC Jimenez and F Jourdan and N Kale and MI Klapa and O Kohlbacher and K Koort and K Kultima and G Le Corguillé and NK Moschonas and S Neumann and C O?Donovan and M Reczko and P Rocca-Serra and A Rosato and RM Salek and SA Sansone and V Satagopam and D Schober and R Shimmo and RA Spicer and O Spjuth and EA Thévenot and MR Viant and RJM Weber and EL Willighagen and G Zanetti and C Steinbeck}, url = {https://doi.org/10.12688/f1000research.12342.2}, year = {2017}, date = {2017-01-01}, journal = {F1000Research}, volume = {6}, number = {1649}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
Flett, Fiona J; Sachsenberg, Timo; Kohlbacher, Oliver; Mackay, Logan C; Interthal, Heidrun Differential Enzymatic 16O/18O Labelling for the Detection of Cross-Linked Nucleic Acid-Protein Heteroconjugates Anal. Chem., 89 (21), pp. 11208-11213, 2017. @article{18OAnalChem2017, title = {Differential Enzymatic 16O/18O Labelling for the Detection of Cross-Linked Nucleic Acid-Protein Heteroconjugates}, author = {Fiona J Flett and Timo Sachsenberg and Oliver Kohlbacher and Logan C Mackay and Heidrun Interthal}, url = {https://doi.org/10.1021/acs.analchem.7b01625}, year = {2017}, date = {2017-01-01}, journal = {Anal. Chem.}, volume = {89}, number = {21}, pages = {11208-11213}, abstract = {Cross-linking of nucleic acids to proteins in combination with mass spectrometry permits the precise identification of interacting residues between nucleic acid-protein complexes. However, the mass spectrometric identification and characterisation of cross-linked nucleic acid-protein heteroconjugates within a complex sample is challenging. Here we establish a novel enzymatic differential 16O/18O labelling approach, which uniquely labels heteroconjugates. We have developed an automated data analysis workflow based on OpenMS for the identification of differentially isotopically labelled heteroconjugates against a complex background. We validated our method using synthetic model DNA oligonucleotide-peptide heteroconjugates which were subjected to the labelling reaction and analysed by high resolution FT-ICR mass spectrometry.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Cross-linking of nucleic acids to proteins in combination with mass spectrometry permits the precise identification of interacting residues between nucleic acid-protein complexes. However, the mass spectrometric identification and characterisation of cross-linked nucleic acid-protein heteroconjugates within a complex sample is challenging. Here we establish a novel enzymatic differential 16O/18O labelling approach, which uniquely labels heteroconjugates. We have developed an automated data analysis workflow based on OpenMS for the identification of differentially isotopically labelled heteroconjugates against a complex background. We validated our method using synthetic model DNA oligonucleotide-peptide heteroconjugates which were subjected to the labelling reaction and analysed by high resolution FT-ICR mass spectrometry. |
Marco, Moreno Di; Schuster, Heiko; Backert, Linus; Ghosh, Michael; Rammensee, Hans-Georg; Stevanovic, Stefan Unveiling the Peptide Motifs of HLA-C and HLA-G from Naturally Presented Peptides and Generation of Binding Prediction Matrices J. Immunol., 199 (8), pp. 2639-2651, 2017. @article{DiMarcoji1700938, title = {Unveiling the Peptide Motifs of HLA-C and HLA-G from Naturally Presented Peptides and Generation of Binding Prediction Matrices}, author = {Moreno Di Marco and Heiko Schuster and Linus Backert and Michael Ghosh and Hans-Georg Rammensee and Stefan Stevanovic}, url = {http://www.jimmunol.org/content/early/2017/09/13/jimmunol.1700938}, year = {2017}, date = {2017-01-01}, journal = {J. Immunol.}, volume = {199}, number = {8}, pages = {2639-2651}, abstract = {The classical HLA-C and the nonclassical HLA-E and HLA-G molecules play important roles both in the innate and adaptive immune system. Starting already during embryogenesis and continuing throughout our lives, these three Ags exert major functions in immune tolerance, defense against infections, and anticancer immune responses. Despite these important roles, identification and characterization of the peptides presented by these molecules has been lacking behind the more abundant HLA-A and HLA-B gene products. In this study, we elucidated the peptide specificities of these HLA molecules using a comprehensive analysis of naturally presented peptides. To that end, the 15 most frequently expressed HLA-C alleles as well as HLA-E*01:01 and HLA-G*01:01 were transfected into lymphoblastoid C1R cells expressing low endogenous HLA. Identification of naturally presented peptides was performed by immunoprecipitation of HLA and subsequent analysis of HLA-bound peptides by liquid chromatographic tandem mass spectrometry. Peptide motifs of HLA-C unveil anchors in position 2 or 3 with high variances between allotypes, and a less variable anchor at the C-terminal end. The previously reported small ligand repertoire of HLA-E was confirmed within our analysis, and we could show that HLA-G combines a large ligand repertoire with distinct features anchoring peptides at positions 3 and 9, supported by an auxiliary anchor in position 1 and preferred residues in positions 2 and 7. The wealth of HLA ligands resulted in prediction matrices for octa-, nona-, and decamers. Matrices were validated in terms of their binding prediction and compared with the latest NetMHC prediction algorithm NetMHCpan-3.0, which demonstrated their predictive power.}, keywords = {}, pubstate = {published}, tppubtype = {article} } The classical HLA-C and the nonclassical HLA-E and HLA-G molecules play important roles both in the innate and adaptive immune system. Starting already during embryogenesis and continuing throughout our lives, these three Ags exert major functions in immune tolerance, defense against infections, and anticancer immune responses. Despite these important roles, identification and characterization of the peptides presented by these molecules has been lacking behind the more abundant HLA-A and HLA-B gene products. In this study, we elucidated the peptide specificities of these HLA molecules using a comprehensive analysis of naturally presented peptides. To that end, the 15 most frequently expressed HLA-C alleles as well as HLA-E*01:01 and HLA-G*01:01 were transfected into lymphoblastoid C1R cells expressing low endogenous HLA. Identification of naturally presented peptides was performed by immunoprecipitation of HLA and subsequent analysis of HLA-bound peptides by liquid chromatographic tandem mass spectrometry. Peptide motifs of HLA-C unveil anchors in position 2 or 3 with high variances between allotypes, and a less variable anchor at the C-terminal end. The previously reported small ligand repertoire of HLA-E was confirmed within our analysis, and we could show that HLA-G combines a large ligand repertoire with distinct features anchoring peptides at positions 3 and 9, supported by an auxiliary anchor in position 1 and preferred residues in positions 2 and 7. The wealth of HLA ligands resulted in prediction matrices for octa-, nona-, and decamers. Matrices were validated in terms of their binding prediction and compared with the latest NetMHC prediction algorithm NetMHCpan-3.0, which demonstrated their predictive power. |
Schuster, Heiko; Peper, Janet Kerstin; Bösmüller, Hans-Christian; Röhle, Kevin; Backert, Linus; Bilich, Tatjana; Ney, Britta; Löffler, Markus W; Kowalewski, Daniel J; Trautwein, Nico; Rabsteyn, Armin; Engler, Tobias; Braun, Sabine; Haen, Sebastian P; Walz, Juliane Sarah; Schmid-Horch, Barbara; Brucker, Sara; Wallwiener, Diethelm; Kohlbacher, Oliver; Fend, Falko; Rammensee, Hans-Georg; Stevanovic, Stefan; Staebler, Annette; Wagner, Philipp The immunopeptidomic landscape of ovarian carcinomas Proc. Natl. Acad. Sci. USA, 114 (46), pp. E9942-E9951, 2017. @article{HLAOvCaPNAS2017, title = {The immunopeptidomic landscape of ovarian carcinomas}, author = {Heiko Schuster and Janet Kerstin Peper and Hans-Christian Bösmüller and Kevin Röhle and Linus Backert and Tatjana Bilich and Britta Ney and Markus W Löffler and Daniel J Kowalewski and Nico Trautwein and Armin Rabsteyn and Tobias Engler and Sabine Braun and Sebastian P Haen and Juliane Sarah Walz and Barbara Schmid-Horch and Sara Brucker and Diethelm Wallwiener and Oliver Kohlbacher and Falko Fend and Hans-Georg Rammensee and Stefan Stevanovic and Annette Staebler and Philipp Wagner}, url = {https://doi.org/10.1073/pnas.1707658114}, year = {2017}, date = {2017-01-01}, journal = {Proc. Natl. Acad. Sci. USA}, volume = {114}, number = {46}, pages = {E9942-E9951}, abstract = {Immunotherapies, particularly checkpoint inhibitors, have set off a revolution in cancer therapy by releasing the power of the immune system. However, only little is known about the antigens that are essentially presented on cancer cells, capable of exposing them to immune cells. Large scale HLA ligandome analysis has enabled us to exhaustively characterize the immunopeptidomic landscape of epithelial ovarian cancers (EOCs). Additional comparative profiling with the immunopeptidome of a variety of benign sources has unveiled a multitude of ovarian cancer antigens (MUC16, MSLN, LGALS1, IDO1, KLK10) to be presented by HLA class I and class II molecules exclusively on ovarian cancer cells. Most strikingly, ligands derived from mucin 16 and mesothelin, a molecular axis of prognostic importance in EOC, are prominent in a majority of patients. Differential gene expression analysis has allowed us to confirm the relevance of these targets for EOC and further provided important insights into the relationship between gene transcript levels and HLA ligand presentation.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Immunotherapies, particularly checkpoint inhibitors, have set off a revolution in cancer therapy by releasing the power of the immune system. However, only little is known about the antigens that are essentially presented on cancer cells, capable of exposing them to immune cells. Large scale HLA ligandome analysis has enabled us to exhaustively characterize the immunopeptidomic landscape of epithelial ovarian cancers (EOCs). Additional comparative profiling with the immunopeptidome of a variety of benign sources has unveiled a multitude of ovarian cancer antigens (MUC16, MSLN, LGALS1, IDO1, KLK10) to be presented by HLA class I and class II molecules exclusively on ovarian cancer cells. Most strikingly, ligands derived from mucin 16 and mesothelin, a molecular axis of prognostic importance in EOC, are prominent in a majority of patients. Differential gene expression analysis has allowed us to confirm the relevance of these targets for EOC and further provided important insights into the relationship between gene transcript levels and HLA ligand presentation. |
Schärfe, Charlotta P I; Tremmel, Roman; Schwab, Matthias; Kohlbacher, Oliver; Marks, Debora S Genetic variation in human drug-related genes Genome Med., 9 , pp. 117, 2017. @article{MyDrugGenomeMed2017, title = {Genetic variation in human drug-related genes}, author = {Charlotta P I Schärfe and Roman Tremmel and Matthias Schwab and Oliver Kohlbacher and Debora S Marks}, url = {http://biorxiv.org/content/early/2017/06/07/147108}, year = {2017}, date = {2017-01-01}, journal = {Genome Med.}, volume = {9}, pages = {117}, abstract = {Background Variability in drug efficacy and adverse effects are observed in clinical practice. While the extent of genetic variability in classic pharmacokinetic genes is rather well understood, the role of genetic variation in drug targets is typically less studied. Methods Based on 60,706 human exomes from the ExAC dataset, we performed an in-depth computational analysis of the prevalence of functional variants in 806 drug-related genes, including 628 known drug targets. We further computed the likelihood of 1236 FDA-approved drugs to be affected by functional variants in their targets in the whole ExAC population as well as different geographic sub-populations. Results We find that most genetic variants in drug-related genes are very rare (f < 0.1%) and thus will likely not be observed in clinical trials. Furthermore, we show that patient risk varies for many drugs and with respect to geographic ancestry. A focused analysis of oncological drug targets indicates that the probability of a patient carrying germline variants in oncological drug targets is, at 44%, high enough to suggest that not only somatic alterations but also germline variants carried over into the tumor genome could affect the response to antineoplastic agents. Conclusions This study indicates that even though many variants are very rare and thus likely not observed in clinical trials, four in five patients are likely to carry a variant with possibly functional effects in a target for commonly prescribed drugs. Such variants could potentially alter drug efficacy.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Background Variability in drug efficacy and adverse effects are observed in clinical practice. While the extent of genetic variability in classic pharmacokinetic genes is rather well understood, the role of genetic variation in drug targets is typically less studied. Methods Based on 60,706 human exomes from the ExAC dataset, we performed an in-depth computational analysis of the prevalence of functional variants in 806 drug-related genes, including 628 known drug targets. We further computed the likelihood of 1236 FDA-approved drugs to be affected by functional variants in their targets in the whole ExAC population as well as different geographic sub-populations. Results We find that most genetic variants in drug-related genes are very rare (f < 0.1%) and thus will likely not be observed in clinical trials. Furthermore, we show that patient risk varies for many drugs and with respect to geographic ancestry. A focused analysis of oncological drug targets indicates that the probability of a patient carrying germline variants in oncological drug targets is, at 44%, high enough to suggest that not only somatic alterations but also germline variants carried over into the tumor genome could affect the response to antineoplastic agents. Conclusions This study indicates that even though many variants are very rare and thus likely not observed in clinical trials, four in five patients are likely to carry a variant with possibly functional effects in a target for commonly prescribed drugs. Such variants could potentially alter drug efficacy. |
Schneider, Lara; Stöckel, Daniel; Kehl, Tim; Gerasch, Andreas; Ludwig, Nicole; Leidinger, Petra; Huwer, Hanno; Tenzer, Stefan; Kohlbacher, Oliver; Hildebrandt, Andreas; Kaufmann, Michael; Gessler, Manfred; Keller, Andreas; Meese, Eckart; Graf, Norbert; Lenhof, Hans-Peter DrugTargetInspector: An assistance tool for patient treatment stratification Int. J. Cancer, 138 (7), pp. 1765-76, 2016. @article{DTI_IJCan2015, title = {DrugTargetInspector: An assistance tool for patient treatment stratification}, author = {Lara Schneider and Daniel Stöckel and Tim Kehl and Andreas Gerasch and Nicole Ludwig and Petra Leidinger and Hanno Huwer and Stefan Tenzer and Oliver Kohlbacher and Andreas Hildebrandt and Michael Kaufmann and Manfred Gessler and Andreas Keller and Eckart Meese and Norbert Graf and Hans-Peter Lenhof}, url = {https://doi.org/10.1002/ijc.29897}, year = {2016}, date = {2016-01-01}, journal = {Int. J. Cancer}, volume = {138}, number = {7}, pages = {1765-76}, abstract = {Cancer is a large class of diseases that are characterized by a common set of features, known as the Hallmarks of Cancer. One of these hallmarks is the acquisition of genome instability and mutations. This, combined with high proliferation rates and failure of repair mechanisms, leads to clonal evolution as well as a high genotypic and phenotypic diversity within the tumor. As a consequence, treatment and therapy of malignant tumors is still a grand challenge. Moreover, under selective pressure, e.g. caused by chemotherapy, resistant subpopulations can emerge that then may lead to relapse. In order to minimize the risk of developing multi-drug resistant tumor cell populations, optimal (combination) therapies have to be determined on the basis of an indepth characterization of the tumor’s genetic and phenotypic makeup, a process which is an important aspect of stratified medicine and precision medicine. We present DrugTargetInspector (DTI), an interactive assistance tool for treatment stratification. DTI analyzes genomic, transcriptomic and proteomic datasets and provides information on deregulated drug targets, enriched biological pathways and deregulated subnetworks, as well as mutations and their potential effects on putative drug targets and genes of interest. To demonstrate DTI’s broad scope of applicability, we present case studies on several cancer types and different types of input - omics data. DTI’s integrative approach allows users to characterize the tumor under investigation based on various -omics datasets and to elucidate putative treatment options based on clinical decision guidelines, but also proposing additional points of intervention that might be neglected otherwise. DTI can be freely accessed at http://dti.bioinf.uni-sb.de.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Cancer is a large class of diseases that are characterized by a common set of features, known as the Hallmarks of Cancer. One of these hallmarks is the acquisition of genome instability and mutations. This, combined with high proliferation rates and failure of repair mechanisms, leads to clonal evolution as well as a high genotypic and phenotypic diversity within the tumor. As a consequence, treatment and therapy of malignant tumors is still a grand challenge. Moreover, under selective pressure, e.g. caused by chemotherapy, resistant subpopulations can emerge that then may lead to relapse. In order to minimize the risk of developing multi-drug resistant tumor cell populations, optimal (combination) therapies have to be determined on the basis of an indepth characterization of the tumor’s genetic and phenotypic makeup, a process which is an important aspect of stratified medicine and precision medicine. We present DrugTargetInspector (DTI), an interactive assistance tool for treatment stratification. DTI analyzes genomic, transcriptomic and proteomic datasets and provides information on deregulated drug targets, enriched biological pathways and deregulated subnetworks, as well as mutations and their potential effects on putative drug targets and genes of interest. To demonstrate DTI’s broad scope of applicability, we present case studies on several cancer types and different types of input - omics data. DTI’s integrative approach allows users to characterize the tumor under investigation based on various -omics datasets and to elucidate putative treatment options based on clinical decision guidelines, but also proposing additional points of intervention that might be neglected otherwise. DTI can be freely accessed at http://dti.bioinf.uni-sb.de. |
Gatto, Laurent; Hansen, Kasper; Hoopmann, Michael; Hermjakob, Henning; Kohlbacher, Oliver; Beyer, Andreas Testing and validation of computational methods for mass spectrometry J. Proteome Res., 15 (3), pp. 809-14, 2016. @article{TestingJPR2016, title = {Testing and validation of computational methods for mass spectrometry}, author = {Laurent Gatto and Kasper Hansen and Michael Hoopmann and Henning Hermjakob and Oliver Kohlbacher and Andreas Beyer}, url = {https://doi.org/10.1021/acs.jproteome.5b00852}, year = {2016}, date = {2016-01-01}, journal = {J. Proteome Res.}, volume = {15}, number = {3}, pages = {809-14}, abstract = {High-throughput methods based on mass spectrometry (proteomics, metabolomics, lipidomics, etc.) produce a wealth of data that cannot be analyzed without computational methods. The impact of the choice of method on the overall result of a biological study is often underappreciated, but different methods can result in very different biological findings. It is thus essential to evaluate and compare the correctness and relative performance of computational methods. The volume of the data as well as the complexity of the algorithms render unbiased comparisons challenging. This paper discusses some problems and challenges in testing and validation of computational methods. We discuss the different types of data (simulated and experimental validation data) as well as different metrics to compare methods. We also introduce a new public repository for mass spectrometric reference data sets (http://compms.org/RefData) that contains a collection of publicly available datasets for performance evaluation for a wide range of different methods.}, keywords = {}, pubstate = {published}, tppubtype = {article} } High-throughput methods based on mass spectrometry (proteomics, metabolomics, lipidomics, etc.) produce a wealth of data that cannot be analyzed without computational methods. The impact of the choice of method on the overall result of a biological study is often underappreciated, but different methods can result in very different biological findings. It is thus essential to evaluate and compare the correctness and relative performance of computational methods. The volume of the data as well as the complexity of the algorithms render unbiased comparisons challenging. This paper discusses some problems and challenges in testing and validation of computational methods. We discuss the different types of data (simulated and experimental validation data) as well as different metrics to compare methods. We also introduce a new public repository for mass spectrometric reference data sets (http://compms.org/RefData) that contains a collection of publicly available datasets for performance evaluation for a wide range of different methods. |
Dammeier, Sascha; Nahnsen, Sven; Veit, Johannes; Wehner, Frank; Ueffing, Marius; Kohlbacher, Oliver Mass Spectrometry-based Proteomics Reveals Organspecific Expression Patterns To Be Used as Forensic Evidence J. Proteome Res., 15 (1), pp. 182-92, 2016. @article{Forensics_JPR_2016, title = {Mass Spectrometry-based Proteomics Reveals Organspecific Expression Patterns To Be Used as Forensic Evidence}, author = {Sascha Dammeier and Sven Nahnsen and Johannes Veit and Frank Wehner and Marius Ueffing and Oliver Kohlbacher}, url = {https://doi.org/10.1021/acs.jproteome.5b00704}, year = {2016}, date = {2016-01-01}, journal = {J. Proteome Res.}, volume = {15}, number = {1}, pages = {182-92}, abstract = {Standard forensic procedures to examine bullets after an exchange of fire include a mechanical or ballistic reconstruction of the event. While this is routine to identify which projectile hit a subject by DNA analysis of biological material on the surface of the projectile, it is rather difficult to determine which projectile caused the lethal injury – often the crucial point with regard to legal proceedings. With respect to fundamental law it is the duty of the public authority to make every endeavor in order to solve every homicide case. To improve forensic examinations we present a forensic proteomic method to investigate biological material from a projectile’s surface and determine the tissues traversed by it. To obtain a range of relevant samples, different major bovine organs were penetrated with projectiles experimentally. After tryptic “on-surface” digestion, mass spectrometry-based proteome analysis and statistical data analysis, we were able to achieve a cross-validated organ classification accuracy of over 99%. Different types of anticipated external variables exhibited no prominent influence on the findings. In addition, shooting experiments were performed to validate the results. Finally, we show that these concepts could be applied to a real case of murder in order to substantially improve the forensic reconstruction.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Standard forensic procedures to examine bullets after an exchange of fire include a mechanical or ballistic reconstruction of the event. While this is routine to identify which projectile hit a subject by DNA analysis of biological material on the surface of the projectile, it is rather difficult to determine which projectile caused the lethal injury – often the crucial point with regard to legal proceedings. With respect to fundamental law it is the duty of the public authority to make every endeavor in order to solve every homicide case. To improve forensic examinations we present a forensic proteomic method to investigate biological material from a projectile’s surface and determine the tissues traversed by it. To obtain a range of relevant samples, different major bovine organs were penetrated with projectiles experimentally. After tryptic “on-surface” digestion, mass spectrometry-based proteome analysis and statistical data analysis, we were able to achieve a cross-validated organ classification accuracy of over 99%. Different types of anticipated external variables exhibited no prominent influence on the findings. In addition, shooting experiments were performed to validate the results. Finally, we show that these concepts could be applied to a real case of murder in order to substantially improve the forensic reconstruction. |
Schubert, Benjamin; Kohlbacher, Oliver Designing string-of-beads vaccines with optimal spacers Genome Med., 8 (1), pp. 9, 2016. @article{SOB-Design2016, title = {Designing string-of-beads vaccines with optimal spacers}, author = {Benjamin Schubert and Oliver Kohlbacher}, url = {http://www.genomemedicine.com/content/8/1/9}, year = {2016}, date = {2016-01-01}, journal = {Genome Med.}, volume = {8}, number = {1}, pages = {9}, abstract = {String-of-beads polypeptides allow convenient delivery of epitope-based vaccines. The success of a polypeptide relies on efficient processing: constituent epitopes need to be recovered while avoiding neo-epitopes from epitope junctions. Spacers between epitopes are employed to ensure this, but spacer selection is non-trivial.We present a framework to determine optimally the length and sequence of a spacer through multi-objective optimization for human leukocyte antigen class I restricted polypeptides. The method yields string-of-bead vaccines with flexible spacer lengths that increase the predicted epitope recovery rate fivefold while reducing the immunogenicity from neo-epitopes by 44 % compared to designs without spacers.}, keywords = {}, pubstate = {published}, tppubtype = {article} } String-of-beads polypeptides allow convenient delivery of epitope-based vaccines. The success of a polypeptide relies on efficient processing: constituent epitopes need to be recovered while avoiding neo-epitopes from epitope junctions. Spacers between epitopes are employed to ensure this, but spacer selection is non-trivial.We present a framework to determine optimally the length and sequence of a spacer through multi-objective optimization for human leukocyte antigen class I restricted polypeptides. The method yields string-of-bead vaccines with flexible spacer lengths that increase the predicted epitope recovery rate fivefold while reducing the immunogenicity from neo-epitopes by 44 % compared to designs without spacers. |
Stickel, Juliane; Kowalewski, Daniel; Walz, Simon; Backert, Linus; Schuster, Heiko; Kohlbacher, Oliver; Weisel, Katja; Rittig, Susanne; Kanz, Lothar; Salih, Helmut; Rammensee, Hans-Georg; Stevanovic, Stefan Carfilzomib alters the HLA-presented peptidome of myeloma cells and impairs presentation of peptides with aromatic C-termini Blood Cancer J., 6 , pp. e411, 2016. @article{BCJStickel2016, title = {Carfilzomib alters the HLA-presented peptidome of myeloma cells and impairs presentation of peptides with aromatic C-termini}, author = {Juliane Stickel and Daniel Kowalewski and Simon Walz and Linus Backert and Heiko Schuster and Oliver Kohlbacher and Katja Weisel and Susanne Rittig and Lothar Kanz and Helmut Salih and Hans-Georg Rammensee and Stefan Stevanovic}, doi = {https://doi.org/10.1038/bcj.2016.14}, year = {2016}, date = {2016-01-01}, journal = {Blood Cancer J.}, volume = {6}, pages = {e411}, abstract = {Recent studies suggest that multiple myeloma is an immunogenic disease, which might be effectively targeted by antigen-specific T-cell immunotherapy. As standard of care in myeloma includes proteasome inhibitor therapy, it is of great importance to characterize the effects of this treatment on HLA-restricted antigen presentation and implement only robustly presented targets for immunotherapeutic intervention. Here, we present a study that longitudinally and semi-quantitatively maps the effects of the proteasome inhibitor carfilzomib on HLA-restricted antigen presentation. The relative presentation levels of 4780 different HLA ligands were quantified in an in vitro model employing carfilzomib treatment of MM.1S and U266 myeloma cells, which revealed significant modulation of a substantial fraction of the HLA-presented peptidome. Strikingly, we detected selective down-modulation of HLA ligands with aromatic C-terminal anchor amino acids. This particularly manifested as a marked reduction in the presentation of HLA ligands through the HLA allotypes A*23:01 and A*24:02 on MM.1S cells. These findings implicate that carfilzomib mediates a direct, peptide motif-specific inhibitory effect on HLA ligand processing and presentation. As a substantial proportion of HLA allotypes present peptides with aromatic C-termini, our results may have broad implications for the implementation of antigen-specific treatment approaches in patients undergoing carfilzomib treatment.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Recent studies suggest that multiple myeloma is an immunogenic disease, which might be effectively targeted by antigen-specific T-cell immunotherapy. As standard of care in myeloma includes proteasome inhibitor therapy, it is of great importance to characterize the effects of this treatment on HLA-restricted antigen presentation and implement only robustly presented targets for immunotherapeutic intervention. Here, we present a study that longitudinally and semi-quantitatively maps the effects of the proteasome inhibitor carfilzomib on HLA-restricted antigen presentation. The relative presentation levels of 4780 different HLA ligands were quantified in an in vitro model employing carfilzomib treatment of MM.1S and U266 myeloma cells, which revealed significant modulation of a substantial fraction of the HLA-presented peptidome. Strikingly, we detected selective down-modulation of HLA ligands with aromatic C-terminal anchor amino acids. This particularly manifested as a marked reduction in the presentation of HLA ligands through the HLA allotypes A*23:01 and A*24:02 on MM.1S cells. These findings implicate that carfilzomib mediates a direct, peptide motif-specific inhibitory effect on HLA ligand processing and presentation. As a substantial proportion of HLA allotypes present peptides with aromatic C-termini, our results may have broad implications for the implementation of antigen-specific treatment approaches in patients undergoing carfilzomib treatment. |
Kohlbacher, Oliver; Vitek, Olga; Weintraub, Susan T Challenges in Large-Scale Computational Mass Spectrometry and Multiomics J. Proteome Res., 15 (3), pp. 681-2, 2016. @article{JPRSIEditorial2016, title = {Challenges in Large-Scale Computational Mass Spectrometry and Multiomics}, author = {Oliver Kohlbacher and Olga Vitek and Susan T Weintraub}, url = {https://pubs.acs.org/doi/abs/10.1021/acs.jproteome.6b00067}, year = {2016}, date = {2016-01-01}, journal = {J. Proteome Res.}, volume = {15}, number = {3}, pages = {681-2}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
Müller, Sabine C; Backes, Christina; Gress, Alexander; Baumgarten, Nina; Kalinina, Olga V; Moll, Andreas; Kohlbacher, Oliver; Meese, Eckart; Keller, Andreas BALL-SNPgp – from genetic variants towards computational diagnostics Bioinformatics, 2016. @article{BALLsnp2016, title = {BALL-SNPgp – from genetic variants towards computational diagnostics}, author = {Sabine C Müller and Christina Backes and Alexander Gress and Nina Baumgarten and Olga V Kalinina and Andreas Moll and Oliver Kohlbacher and Eckart Meese and Andreas Keller}, doi = {https://doi.org/10.1093/bioinformatics/btw084}, year = {2016}, date = {2016-01-01}, journal = {Bioinformatics}, abstract = {Summary: In medical research, it is crucial to understand the functional consequences of genetic alterations, for example non-synonymous single nucleotide variants (nsSNVs). NsSNVs are known to be causative for several human diseases. However, the genetic basis of complex disorders such as diabetes or cancer comprises multiple factors. Methods to analyze putative synergetic effects of multiple such factors, however, are limited. Here, we concentrate on nsSNVs and present BALL-SNPgp, a tool for structural and functional characterization of nsSNVs, which is aimed to improve pathogenicity assessment in computational diagnostics. Based on annotated SNV data, BALL-SNPgp creates a 3D visualization of the encoded protein, collects available information from different resources concerning disease relevance and other functional annotations, performs cluster analysis, predicts putative binding pockets and provides data on known interaction sites. Availability and implementation: BALL-SNPgp is based on the comprehensive C++ framework Biochemical Algorithms Library (BALL) and its visualization front-end BALLView. Our tool is available at www.ccb.uni-saarland.de/BALL-SNPgp.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Summary: In medical research, it is crucial to understand the functional consequences of genetic alterations, for example non-synonymous single nucleotide variants (nsSNVs). NsSNVs are known to be causative for several human diseases. However, the genetic basis of complex disorders such as diabetes or cancer comprises multiple factors. Methods to analyze putative synergetic effects of multiple such factors, however, are limited. Here, we concentrate on nsSNVs and present BALL-SNPgp, a tool for structural and functional characterization of nsSNVs, which is aimed to improve pathogenicity assessment in computational diagnostics. Based on annotated SNV data, BALL-SNPgp creates a 3D visualization of the encoded protein, collects available information from different resources concerning disease relevance and other functional annotations, performs cluster analysis, predicts putative binding pockets and provides data on known interaction sites. Availability and implementation: BALL-SNPgp is based on the comprehensive C++ framework Biochemical Algorithms Library (BALL) and its visualization front-end BALLView. Our tool is available at www.ccb.uni-saarland.de/BALL-SNPgp. |
Schubert, Benjamin; Walzer, Mathias; Brachvogel, Hans-Philipp; Szolek, Andras; Mohr, Christopher; Kohlbacher, Oliver FRED 2 – An Immunoinformatics Framework for Python Bioinformatics, 32 (13), pp. 2044-6, 2016. @article{FRED2_Bioinfo, title = {FRED 2 – An Immunoinformatics Framework for Python}, author = {Benjamin Schubert and Mathias Walzer and Hans-Philipp Brachvogel and Andras Szolek and Christopher Mohr and Oliver Kohlbacher}, doi = {https://doi.org/10.1093/bioinformatics/btw113}, year = {2016}, date = {2016-01-01}, journal = {Bioinformatics}, volume = {32}, number = {13}, pages = {2044-6}, abstract = {Summary: Immunoinformatics approaches are widely used in a variety of applications from basic immunological to applied biomedical research. Complex data integration is inevitable in immunological research and usually requires comprehensive pipelines including multiple tools and data sources. Non-standard input and output formats of immunoinformatics tools make the development of such applications difficult. Here we present FRED 2, an open-source immunoinformatics framework offering easy and unified access to methods for epitope prediction and other immunoinformatics applications. FRED 2 is implemented in Python and designed to be extendable and flexible to allow rapid prototyping of complex applications. Availability: FRED 2 is available at http://fred-2.github.io Contact: Supplementary information: Supplementary data are available at Bioinformatics online.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Summary: Immunoinformatics approaches are widely used in a variety of applications from basic immunological to applied biomedical research. Complex data integration is inevitable in immunological research and usually requires comprehensive pipelines including multiple tools and data sources. Non-standard input and output formats of immunoinformatics tools make the development of such applications difficult. Here we present FRED 2, an open-source immunoinformatics framework offering easy and unified access to methods for epitope prediction and other immunoinformatics applications. FRED 2 is implemented in Python and designed to be extendable and flexible to allow rapid prototyping of complex applications. Availability: FRED 2 is available at http://fred-2.github.io Contact: Supplementary information: Supplementary data are available at Bioinformatics online. |
de la Garza, Luis; Veit, Johannes; Szolek, Andras; Röttig, Marc; Aiche, Stephan; Gesing, Sandra; Reinert, Knut; Kohlbacher, Oliver From the Desktop to the Grid: scalable Bioinformatics via Workflow Conversion BMC Bioinformatics, 17 (1), pp. 127, 2016. @article{delaGarza2016, title = {From the Desktop to the Grid: scalable Bioinformatics via Workflow Conversion}, author = {Luis de la Garza and Johannes Veit and Andras Szolek and Marc Röttig and Stephan Aiche and Sandra Gesing and Knut Reinert and Oliver Kohlbacher}, doi = {https://doi.org/10.1186/s12859-016-0978-9}, year = {2016}, date = {2016-01-01}, journal = {BMC Bioinformatics}, volume = {17}, number = {1}, pages = {127}, abstract = {BACKGROUND: Reproducibility is one of the tenets of the scientific method. Scientific experiments often comprise complex data flows, selection of adequate parameters, and analysis and visualization of intermediate and end results. Breaking down the complexity of such experiments into the joint collaboration of small, repeatable, well defined tasks, each with well defined inputs, parameters, and outputs, offers the immediate benefit of identifying bottlenecks, pinpoint sections which could benefit from parallelization, among others. Workflows rest upon the notion of splitting complex work into the joint effort of several manageable tasks. There are several engines that give users the ability to design and execute workflows. Each engine was created to address certain problems of a specific community, therefore each one has its advantages and shortcomings. Furthermore, not all features of all workflow engines are royalty-free -an aspect that could potentially drive away members of the scientific community. RESULTS: We have developed a set of tools that enables the scientific community to benefit from workflow interoperability. We developed a platform-free structured representation of parameters, inputs, outputs of command-line tools in so-called Common Tool Descriptor documents. We have also overcome the shortcomings and combined the features of two royalty-free workflow engines with a substantial user community: the Konstanz Information Miner, an engine which we see as a formidable workflow editor, and the Grid and User Support Environment, a web-based framework able to interact with several high-performance computing resources. We have thus created a free and highly accessible way to design workflows on a desktop computer and execute them on high-performance computing resources. CONCLUSIONS: Our work will not only reduce time spent on designing scientific workflows, but also make executing workflows on remote high-performance computing resources more accessible to technically inexperienced users. We strongly believe that our efforts not only decrease the turnaround time to obtain scientific results but also have a positive impact on reproducibility, thus elevating the quality of obtained scientific results.}, keywords = {}, pubstate = {published}, tppubtype = {article} } BACKGROUND: Reproducibility is one of the tenets of the scientific method. Scientific experiments often comprise complex data flows, selection of adequate parameters, and analysis and visualization of intermediate and end results. Breaking down the complexity of such experiments into the joint collaboration of small, repeatable, well defined tasks, each with well defined inputs, parameters, and outputs, offers the immediate benefit of identifying bottlenecks, pinpoint sections which could benefit from parallelization, among others. Workflows rest upon the notion of splitting complex work into the joint effort of several manageable tasks. There are several engines that give users the ability to design and execute workflows. Each engine was created to address certain problems of a specific community, therefore each one has its advantages and shortcomings. Furthermore, not all features of all workflow engines are royalty-free -an aspect that could potentially drive away members of the scientific community. RESULTS: We have developed a set of tools that enables the scientific community to benefit from workflow interoperability. We developed a platform-free structured representation of parameters, inputs, outputs of command-line tools in so-called Common Tool Descriptor documents. We have also overcome the shortcomings and combined the features of two royalty-free workflow engines with a substantial user community: the Konstanz Information Miner, an engine which we see as a formidable workflow editor, and the Grid and User Support Environment, a web-based framework able to interact with several high-performance computing resources. We have thus created a free and highly accessible way to design workflows on a desktop computer and execute them on high-performance computing resources. CONCLUSIONS: Our work will not only reduce time spent on designing scientific workflows, but also make executing workflows on remote high-performance computing resources more accessible to technically inexperienced users. We strongly believe that our efforts not only decrease the turnaround time to obtain scientific results but also have a positive impact on reproducibility, thus elevating the quality of obtained scientific results. |
Breckels, Lisa M; Holden, Sean; Wojnar, David; Mulvey, Claire M; Christoforou, Adny; Groen, Arnoud; Kohlbacher, Oliver; Lilley, Kathryn S; Gatto, Laurent Learning from Heterogeneous Data Sources: An Application in Spatial Proteomics PLoS Comput. Biol., 12 (5), pp. e1004920, 2016. @article{SpatialProtPLOSCB, title = {Learning from Heterogeneous Data Sources: An Application in Spatial Proteomics}, author = {Lisa M Breckels and Sean Holden and David Wojnar and Claire M Mulvey and Adny Christoforou and Arnoud Groen and Oliver Kohlbacher and Kathryn S Lilley and Laurent Gatto}, doi = {https://doi.org/10.1371/journal.pcbi.1004920}, year = {2016}, date = {2016-01-01}, journal = {PLoS Comput. Biol.}, volume = {12}, number = {5}, pages = {e1004920}, abstract = {Sub-cellular localisation of proteins is an essential post-translational regulatory mechanism that can be assayed using high-throughput mass spectrometry (MS). These MS-based spatial proteomics experiments enable us to pinpoint the sub-cellular distribution of thousands of proteins in a specific system under controlled conditions. Recent advances in high-throughput MS methods have yielded a plethora of experimental spatial proteomics data for the cell biology community. Yet, there are many third-party data sources, such as immunofluorescence microscopy or protein annotations and sequences, which represent a rich and vast source of complementary information. We present a unique transfer learning classification framework that utilises a nearest-neighbour or support vector machine system, to integrate heterogeneous data sources to considerably improve on the quantity and quality of sub-cellular protein assignment. We demonstrate the utility of our algorithms through evaluation of five experimental datasets, from four different species in conjunction with four different auxiliary data sources to classify proteins to tens of sub-cellular compartments with high generalisation accuracy. We further apply the method to an experiment on pluripotent mouse embryonic stem cells to classify a set of previously unknown proteins, and validate our findings against a recent high resolution map of the mouse stem cell proteome. The methodology is distributed as part of the open-source Bioconductor pRoloc suite for spatial proteomics data analysis.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Sub-cellular localisation of proteins is an essential post-translational regulatory mechanism that can be assayed using high-throughput mass spectrometry (MS). These MS-based spatial proteomics experiments enable us to pinpoint the sub-cellular distribution of thousands of proteins in a specific system under controlled conditions. Recent advances in high-throughput MS methods have yielded a plethora of experimental spatial proteomics data for the cell biology community. Yet, there are many third-party data sources, such as immunofluorescence microscopy or protein annotations and sequences, which represent a rich and vast source of complementary information. We present a unique transfer learning classification framework that utilises a nearest-neighbour or support vector machine system, to integrate heterogeneous data sources to considerably improve on the quantity and quality of sub-cellular protein assignment. We demonstrate the utility of our algorithms through evaluation of five experimental datasets, from four different species in conjunction with four different auxiliary data sources to classify proteins to tens of sub-cellular compartments with high generalisation accuracy. We further apply the method to an experiment on pluripotent mouse embryonic stem cells to classify a set of previously unknown proteins, and validate our findings against a recent high resolution map of the mouse stem cell proteome. The methodology is distributed as part of the open-source Bioconductor pRoloc suite for spatial proteomics data analysis. |
Mohr, Christopher; Friedrich, Andreas; Wojnar, David; Kenar, Erhan; Polatkan, Aydin Can; Codrea, Marius Cosmin; Czemmel, Stefan; Kohlbacher, Oliver; Nahnsen, Sven qPortal – A Science Gateway for Biomedical Applications Proc. Int. Workshop Science Gateways (IWSG 2016), 2016. (BibTeX) @article{qPortal-IWSG2016, title = {qPortal – A Science Gateway for Biomedical Applications}, author = {Christopher Mohr and Andreas Friedrich and David Wojnar and Erhan Kenar and Aydin Can Polatkan and Marius Cosmin Codrea and Stefan Czemmel and Oliver Kohlbacher and Sven Nahnsen}, year = {2016}, date = {2016-01-01}, journal = {Proc. Int. Workshop Science Gateways (IWSG 2016)}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
de la Garza, Luis; Aicheler, Fabian; Kohlbacher, Oliver From the Desktop to the Grid and Cloud: Conversion of KNIME Workflows to gUSE Proc. Int. Workshop Science Gateways (IWSG 2016), 2016. (BibTeX) @article{KNIME2gUSE-IWSG2016, title = {From the Desktop to the Grid and Cloud: Conversion of KNIME Workflows to gUSE}, author = {Luis de la Garza and Fabian Aicheler and Oliver Kohlbacher}, year = {2016}, date = {2016-01-01}, journal = {Proc. Int. Workshop Science Gateways (IWSG 2016)}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
Griss, Johannes; Perez-Riverol, Yasset; Lewis, Steve; Tabb, David L; Dianes, José A; del-Toro, Noemi; Rurik, Marc; Walzer, Mathias M; Kohlbacher, Oliver; Hermjakob, Henning; Wang, Rui; Vizcaíno, Juan Antonio Recognizing millions of consistently unidentified spectra across hundreds of shotgun proteomics datasets Nat. Methods, 13 (8), pp. 651-656, 2016. @article{ClusteringNatMethods2016, title = {Recognizing millions of consistently unidentified spectra across hundreds of shotgun proteomics datasets}, author = {Johannes Griss and Yasset Perez-Riverol and Steve Lewis and David L Tabb and José A Dianes and Noemi del-Toro and Marc Rurik and Mathias M Walzer and Oliver Kohlbacher and Henning Hermjakob and Rui Wang and Juan Antonio Vizcaíno}, doi = {https://dx.doi.org/10.1038%2Fnmeth.3902}, year = {2016}, date = {2016-01-01}, journal = {Nat. Methods}, volume = {13}, number = {8}, pages = {651-656}, abstract = {Mass spectrometry (MS) is the main technology used in proteomics approaches. However, on average, 75% of spectra analyzed in an MS experiment remain unidentified. We propose to use spectrum clustering at a large scale to shed light on these unidentified spectra. The Proteomics Identifications (PRIDE) Database Archive is one of the largest MS proteomics public data repositories worldwide. By clustering all tandem MS spectra publicly available in the PRIDE Archive, coming from hundreds of data sets, we were able to consistently characterize spectra into three distinct groups: (1) incorrectly identified, (2) correctly identified but below the set scoring threshold, and (3) truly unidentified. Using multiple complementary analysis approaches, we were able to identify ~20% of the consistently unidentified spectra. The complete spectrum-clustering results are available through the new version of the PRIDE Cluster resource (http://www.ebi.ac.uk/pride/cluster). This resource is intended, among other aims, to encourage and simplify further investigation into these unidentified spectra.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Mass spectrometry (MS) is the main technology used in proteomics approaches. However, on average, 75% of spectra analyzed in an MS experiment remain unidentified. We propose to use spectrum clustering at a large scale to shed light on these unidentified spectra. The Proteomics Identifications (PRIDE) Database Archive is one of the largest MS proteomics public data repositories worldwide. By clustering all tandem MS spectra publicly available in the PRIDE Archive, coming from hundreds of data sets, we were able to consistently characterize spectra into three distinct groups: (1) incorrectly identified, (2) correctly identified but below the set scoring threshold, and (3) truly unidentified. Using multiple complementary analysis approaches, we were able to identify ~20% of the consistently unidentified spectra. The complete spectrum-clustering results are available through the new version of the PRIDE Cluster resource (http://www.ebi.ac.uk/pride/cluster). This resource is intended, among other aims, to encourage and simplify further investigation into these unidentified spectra. |
Ranninger, Christina; Schmidt, Lukas E; Rurik, Marc; Limonciel, Alice; Jennings, Paul; Kohlbacher, Oliver; Huber, Christian G Improving global feature detectabilities through scan range splitting for untargeted metabolomics by high-performance liquid chromatography-Orbitrap mass spectrometry Anal. Chim. Acta, 930 , pp. 13-22, 2016. @article{ScanRangeACA2016, title = {Improving global feature detectabilities through scan range splitting for untargeted metabolomics by high-performance liquid chromatography-Orbitrap mass spectrometry}, author = {Christina Ranninger and Lukas E Schmidt and Marc Rurik and Alice Limonciel and Paul Jennings and Oliver Kohlbacher and Christian G Huber}, doi = {https://doi.org/10.1016/j.aca.2016.05.017}, year = {2016}, date = {2016-01-01}, journal = {Anal. Chim. Acta}, volume = {930}, pages = {13-22}, abstract = {Untargeted metabolomics aims at obtaining quantitative information on the highest possible number of low-molecular biomolecules present in a biological sample. Rather small changes in mass spectrometric spectrum acquisition parameters may have a significant influence on the detectabilities of metabolites in untargeted global-scale studies by means of high-performance liquid chromatography-mass spectrometry (HPLC-MS). Employing whole cell lysates of human renal proximal tubule cells, we present a systematic global-scale study of the influence of mass spectrometric scan parameters and post-acquisition data treatment on the number and intensity of metabolites detectable in whole cell lysates. Ion transmission and ion collection efficiencies in an Orbitrap-based mass spectrometer basically depend on the m/z range scanned, which, ideally, requires different instrument settings for the respective mass ranges investigated. Therefore, we split a full scan range of m/z 50-1000 relevant for metabolites into two separate segments (m/z 50-200 and m/z 200-1,000), allowing an independent tuning of the ion transmission parameters for both mass ranges. Three different implementations, involving either scanning from m/z 50-1000 in a single scan, or scanning from m/z 50-200 and from m/z 200-1000 in two alternating scans, or performing two separate HPLC-MS runs with m/z 50-200 and m/z 200-1000 scan ranges were critically assessed. The detected features were subjected to rigorous background filtering and quality control in order to obtain reliable metabolite features for subsequent differential quantification. The most efficient approach in terms of feature number, which forms the basis for statistical analysis, identification, and for generating biological hypotheses, was the separate analysis of two different mass ranges. This lead to an increase in the number of detectable metabolite features, especially in the higher mass range (m/z greater than 400), by 2.5 (negative mode) to 6-fold (positive mode) as compared to analysis involving a single scan range. The total number of features confidently detectable was 560 in positive ion mode, and 436 in negative ion mode.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Untargeted metabolomics aims at obtaining quantitative information on the highest possible number of low-molecular biomolecules present in a biological sample. Rather small changes in mass spectrometric spectrum acquisition parameters may have a significant influence on the detectabilities of metabolites in untargeted global-scale studies by means of high-performance liquid chromatography-mass spectrometry (HPLC-MS). Employing whole cell lysates of human renal proximal tubule cells, we present a systematic global-scale study of the influence of mass spectrometric scan parameters and post-acquisition data treatment on the number and intensity of metabolites detectable in whole cell lysates. Ion transmission and ion collection efficiencies in an Orbitrap-based mass spectrometer basically depend on the m/z range scanned, which, ideally, requires different instrument settings for the respective mass ranges investigated. Therefore, we split a full scan range of m/z 50-1000 relevant for metabolites into two separate segments (m/z 50-200 and m/z 200-1,000), allowing an independent tuning of the ion transmission parameters for both mass ranges. Three different implementations, involving either scanning from m/z 50-1000 in a single scan, or scanning from m/z 50-200 and from m/z 200-1000 in two alternating scans, or performing two separate HPLC-MS runs with m/z 50-200 and m/z 200-1000 scan ranges were critically assessed. The detected features were subjected to rigorous background filtering and quality control in order to obtain reliable metabolite features for subsequent differential quantification. The most efficient approach in terms of feature number, which forms the basis for statistical analysis, identification, and for generating biological hypotheses, was the separate analysis of two different mass ranges. This lead to an increase in the number of detectable metabolite features, especially in the higher mass range (m/z greater than 400), by 2.5 (negative mode) to 6-fold (positive mode) as compared to analysis involving a single scan range. The total number of features confidently detectable was 560 in positive ion mode, and 436 in negative ion mode. |
Perez-Riverol, Yasset; Gatto, Laurent; Wang, Rui; Sachsenberg, Timo; Uszkoreit, Julian; Leprevost, Felipe; Fufezan, Christian; Ternent, Tobias; Eglen, Stephen J; Katz, Daniel S S; Pollard, Tom J; Konovalov, Alexander; Flight, Robert M; Blin, Kai; Vizcaino, Juan Antonio Ten Simple Rules for Taking Advantage of git and GitHub bioRxiv, 2016. @article{Perez-Riverol048744, title = {Ten Simple Rules for Taking Advantage of git and GitHub}, author = {Yasset Perez-Riverol and Laurent Gatto and Rui Wang and Timo Sachsenberg and Julian Uszkoreit and Felipe Leprevost and Christian Fufezan and Tobias Ternent and Stephen J Eglen and Daniel S S Katz and Tom J Pollard and Alexander Konovalov and Robert M Flight and Kai Blin and Juan Antonio Vizcaino}, url = {http://www.biorxiv.org/content/early/2016/05/13/048744}, year = {2016}, date = {2016-01-01}, journal = {bioRxiv}, abstract = {A ’Ten Simple Rules’ guide to git and GitHub. We describe and provide examples on how to use these software to track projects, as users, teams and organizations. We document collaborative development using branching and forking, interaction between collaborators using issues and continuous integration and automation using, for example, Travis CI and codecov. We also describe dissemination and social aspects of GitHub such as GitHub pages, following and watching repositories, and give advice on how to make code citable.}, keywords = {}, pubstate = {published}, tppubtype = {article} } A ’Ten Simple Rules’ guide to git and GitHub. We describe and provide examples on how to use these software to track projects, as users, teams and organizations. We document collaborative development using branching and forking, interaction between collaborators using issues and continuous integration and automation using, for example, Travis CI and codecov. We also describe dissemination and social aspects of GitHub such as GitHub pages, following and watching repositories, and give advice on how to make code citable. |
Löffler, Markus; Chandra, Anoop P; Laske, Karoline; Schroeder, Christopher; Bonzheim, Irina; Hilke, Franz J; Kowalewski, Daniel J; Trautwein, Nico; Schuster, Heiko; Gründer, Marc; Walzer, Mathias; Mohr, Christopher; Nguyen, Huu-Phuc; Riess, Olaf; Bauer, Peter; Nahnsen, Sven; Königsrainer, Alfred; Nadalina, Silvio; Zieker, Derek; Glatzle, Jörg; Thiel, Karolin; Clasen, Stephan; Bösmüller, Hans; Fend, Falko; Kohlbacher, Oliver; Gouttefangeas, Cecile; Stevanovic, Stefan; Rammensee, Hans-Georg Personalized peptide vaccine induced immune response associated with long-term survival of a metastatic cholangiocarcinoma patient J. Hepatol., 65 (4), pp. 849-55, 2016. @article{CaseStudyJHepatol2016, title = {Personalized peptide vaccine induced immune response associated with long-term survival of a metastatic cholangiocarcinoma patient}, author = {Markus Löffler and Anoop P Chandra and Karoline Laske and Christopher Schroeder and Irina Bonzheim and Franz J Hilke and Daniel J Kowalewski and Nico Trautwein and Heiko Schuster and Marc Gründer and Mathias Walzer and Christopher Mohr and Huu-Phuc Nguyen and Olaf Riess and Peter Bauer and Sven Nahnsen and Alfred Königsrainer and Silvio Nadalina and Derek Zieker and Jörg Glatzle and Karolin Thiel and Stephan Clasen and Hans Bösmüller and Falko Fend and Oliver Kohlbacher and Cecile Gouttefangeas and Stefan Stevanovic and Hans-Georg Rammensee}, doi = {https://doi.org/10.1016/j.jhep.2016.06.027}, year = {2016}, date = {2016-01-01}, journal = {J. Hepatol.}, volume = {65}, number = {4}, pages = {849-55}, abstract = {BACKGROUND & AIMS: We report a novel experimental immunotherapeutic approach in a patient with metastatic intrahepatic cholangiocarcinoma. In the 5year course of the disease, the initial tumor mass, two local recurrences and a lung metastasis were surgically removed. Lacking alternative treatment options, aiming at the induction of anti-tumor T cells responses, we initiated a personalized multi-peptide vaccination, based on in-depth analysis of tumor antigens (immunopeptidome) and sequencing. METHODS: Tumors were characterized by immunohistochemistry, next-generation sequencing and mass spectrometry of HLA ligands. RESULTS: Although several tumor-specific neo-epitopes were predicted in silico, none could be validated by mass spectrometry. Instead, a personalized multi-peptide vaccine containing non-mutated tumor-associated epitopes was designed and applied. Immunomonitoring showed vaccine-induced T cell responses to three out of seven peptides administered. The pulmonary metastasis resected after start of vaccination showed strong immune cell infiltration and perforin positivity, in contrast to the previous lesions. The patient remains clinically healthy, without any radiologically detectable tumors since March 2013 and the vaccination is continued. CONCLUSIONS: This remarkable clinical course encourages formal clinical studies on adjuvant personalized peptide vaccination in cholangiocarcinoma. LAY SUMMARY: Metastatic cholangiocarcinomas, cancers that originate from the liver bile ducts, have very limited treatment options and a fatal prognosis. We describe a novel therapeutic approach in such a patient using a personalized multi-peptide vaccine. This vaccine, developed based on the characterization of the patient's tumor, evoked detectable anti-tumor immune responses, associating with long-term tumor-free survival.}, keywords = {}, pubstate = {published}, tppubtype = {article} } BACKGROUND & AIMS: We report a novel experimental immunotherapeutic approach in a patient with metastatic intrahepatic cholangiocarcinoma. In the 5year course of the disease, the initial tumor mass, two local recurrences and a lung metastasis were surgically removed. Lacking alternative treatment options, aiming at the induction of anti-tumor T cells responses, we initiated a personalized multi-peptide vaccination, based on in-depth analysis of tumor antigens (immunopeptidome) and sequencing. METHODS: Tumors were characterized by immunohistochemistry, next-generation sequencing and mass spectrometry of HLA ligands. RESULTS: Although several tumor-specific neo-epitopes were predicted in silico, none could be validated by mass spectrometry. Instead, a personalized multi-peptide vaccine containing non-mutated tumor-associated epitopes was designed and applied. Immunomonitoring showed vaccine-induced T cell responses to three out of seven peptides administered. The pulmonary metastasis resected after start of vaccination showed strong immune cell infiltration and perforin positivity, in contrast to the previous lesions. The patient remains clinically healthy, without any radiologically detectable tumors since March 2013 and the vaccination is continued. CONCLUSIONS: This remarkable clinical course encourages formal clinical studies on adjuvant personalized peptide vaccination in cholangiocarcinoma. LAY SUMMARY: Metastatic cholangiocarcinomas, cancers that originate from the liver bile ducts, have very limited treatment options and a fatal prognosis. We describe a novel therapeutic approach in such a patient using a personalized multi-peptide vaccine. This vaccine, developed based on the characterization of the patient's tumor, evoked detectable anti-tumor immune responses, associating with long-term tumor-free survival. |
Röst, Hannes L; Sachsenberg, Timo; Aiche, Stephan; Bielow, Chris; Weisser, Hendrik; Aicheler, Fabian; Andreotti, Sandro; Ehrlich, Hans-Christian; Gutenbrunner, Petra; Kenar, Erhan; Liang, Xiao; Nahnsen, Sven; Nilse, Lars; Pfeuffer, Julianus; Rosenberger, George; Rurik, Marc; Schmitt, Uwe; Veit, Johannes; Walzer, Mathias; Wojnar, David; Wolski, Witold E; Schilling, Oliver; Choudhary, Jyoti S; Malmström, Lars; Aebersold, Ruedi; Reinert, Knut; Kohlbacher, Oliver OpenMS: A flexible open-source software platform for mass spectrometry data analysis Nat. Methods, 13 (9), pp. 741-748, 2016. @article{OpenMS2_2016, title = {OpenMS: A flexible open-source software platform for mass spectrometry data analysis}, author = {Hannes L Röst and Timo Sachsenberg and Stephan Aiche and Chris Bielow and Hendrik Weisser and Fabian Aicheler and Sandro Andreotti and Hans-Christian Ehrlich and Petra Gutenbrunner and Erhan Kenar and Xiao Liang and Sven Nahnsen and Lars Nilse and Julianus Pfeuffer and George Rosenberger and Marc Rurik and Uwe Schmitt and Johannes Veit and Mathias Walzer and David Wojnar and Witold E Wolski and Oliver Schilling and Jyoti S Choudhary and Lars Malmström and Ruedi Aebersold and Knut Reinert and Oliver Kohlbacher}, doi = {https://doi.org/10.1038/nmeth.3959}, year = {2016}, date = {2016-01-01}, journal = {Nat. Methods}, volume = {13}, number = {9}, pages = {741-748}, abstract = {High-resolution mass spectrometry (MS) has become an important tool in the life sciences, contributing to the diagnosis and understanding of human diseases, elucidating structural information and characterizing cellular signaling networks. However, the rapid growth in the volume and complexity of mass spectrometry data makes transparent, accurate and reproducible analysis difficult. We present OpenMS 2.0 (http://www.openms.org/), a robust open-source, cross-platform software specifically designed for the flexible and reproducible analysis of high-throughput MS data. The extensible OpenMS software implements common mass spectrometric data processing tasks through a well-defined API in C++ and Python using standardized open data formats. OpenMS additionally provides a set of 185 tools and ready-made workflow to carry out common mass spectrometric data processing tasks, which enable users to perform complex quantitative mass spectrometric analyses with ease.}, keywords = {}, pubstate = {published}, tppubtype = {article} } High-resolution mass spectrometry (MS) has become an important tool in the life sciences, contributing to the diagnosis and understanding of human diseases, elucidating structural information and characterizing cellular signaling networks. However, the rapid growth in the volume and complexity of mass spectrometry data makes transparent, accurate and reproducible analysis difficult. We present OpenMS 2.0 (http://www.openms.org/), a robust open-source, cross-platform software specifically designed for the flexible and reproducible analysis of high-throughput MS data. The extensible OpenMS software implements common mass spectrometric data processing tasks through a well-defined API in C++ and Python using standardized open data formats. OpenMS additionally provides a set of 185 tools and ready-made workflow to carry out common mass spectrometric data processing tasks, which enable users to perform complex quantitative mass spectrometric analyses with ease. |
Veit, Johannes; Sachsenberg, Timo; Chernev, Alexsandr; Aicheler, Fabian; Urlaub, Henning; Kohlbacher, Oliver LFQProfiler and RNPxl: Open-Source Tools for Label-Free Quantification and Protein-RNA Cross-Linking Integrated into Proteome Discoverer J. Proteome Res., 15 (9), pp. 3441-8., 2016. @article{RNPxlPD_JPR_2016, title = {LFQProfiler and RNPxl: Open-Source Tools for Label-Free Quantification and Protein-RNA Cross-Linking Integrated into Proteome Discoverer}, author = {Johannes Veit and Timo Sachsenberg and Alexsandr Chernev and Fabian Aicheler and Henning Urlaub and Oliver Kohlbacher}, url = {https://pubs.acs.org/doi/abs/10.1021/acs.jproteome.6b00407}, year = {2016}, date = {2016-01-01}, journal = {J. Proteome Res.}, volume = {15}, number = {9}, pages = {3441-8.}, abstract = {Modern mass spectrometry setups used in today's proteomics studies generate vast amounts of raw data, calling for highly efficient data processing and analysis tools. Software for analyzing these data is either monolithic (easy to use, but sometimes too rigid) or workflow-driven (easy to customize, but sometimes complex). Thermo Proteome Discoverer (PD) is a powerful software for workflow-driven data analysis in proteomics which, in our eyes, achieves a good trade-off between flexibility and usability. Here, we present two open-source plugins for PD providing additional functionality: LFQProfiler for label-free quantification of peptides and proteins, and RNPxl for UV-induced peptide-RNA cross-linking data analysis. LFQProfiler interacts with existing PD nodes for peptide identification and validation and takes care of the entire quantitative part of the workflow. We show that it performs at least on par with other state-of-the-art software solutions for label-free quantification in a recently published benchmark ( Ramus, C.; J. Proteomics 2016 , 132 , 51 - 62 ). The second workflow, RNPxl, represents the first software solution to date for identification of peptide-RNA cross-links including automatic localization of the cross-links at amino acid resolution and localization scoring. It comes with a customized integrated cross-link fragment spectrum viewer for convenient manual inspection and validation of the results.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Modern mass spectrometry setups used in today's proteomics studies generate vast amounts of raw data, calling for highly efficient data processing and analysis tools. Software for analyzing these data is either monolithic (easy to use, but sometimes too rigid) or workflow-driven (easy to customize, but sometimes complex). Thermo Proteome Discoverer (PD) is a powerful software for workflow-driven data analysis in proteomics which, in our eyes, achieves a good trade-off between flexibility and usability. Here, we present two open-source plugins for PD providing additional functionality: LFQProfiler for label-free quantification of peptides and proteins, and RNPxl for UV-induced peptide-RNA cross-linking data analysis. LFQProfiler interacts with existing PD nodes for peptide identification and validation and takes care of the entire quantitative part of the workflow. We show that it performs at least on par with other state-of-the-art software solutions for label-free quantification in a recently published benchmark ( Ramus, C.; J. Proteomics 2016 , 132 , 51 - 62 ). The second workflow, RNPxl, represents the first software solution to date for identification of peptide-RNA cross-links including automatic localization of the cross-links at amino acid resolution and localization scoring. It comes with a customized integrated cross-link fragment spectrum viewer for convenient manual inspection and validation of the results. |
Boyles, Matthe; Ranninger, Christina; Reischl, Roland; Rurik, Marc; Tessadri, Richard; Kohlbacher, Oliver; Duschl, Albert; Huber, Christian Copper oxide nanoparticle toxicity profiling using untargeted metabolomics Part. Fiber Toxicol., 13 , pp. 49, 2016. @article{CuOTox2016, title = {Copper oxide nanoparticle toxicity profiling using untargeted metabolomics}, author = {Matthe Boyles and Christina Ranninger and Roland Reischl and Marc Rurik and Richard Tessadri and Oliver Kohlbacher and Albert Duschl and Christian Huber}, doi = {https://doi.org/10.1186/s12989-016-0160-6}, year = {2016}, date = {2016-01-01}, journal = {Part. Fiber Toxicol.}, volume = {13}, pages = {49}, abstract = {Background The rapidly increasing number of engineered nanoparticles (NPs), and products containing NPs, raises concerns for human exposure and safety. With this increasing, and ever changing, catalogue of NPs it is becoming more difficult to adequately assess the toxic potential of new materials in a timely fashion. It is therefore important to develop methods which can provide high-throughput screening of biological responses. The use of omics technologies, including metabolomics, can play a vital role in this process by providing relatively fast, comprehensive, and cost-effective assessment of cellular responses. These techniques thus provide the opportunity to identify specific toxicity pathways and to generate hypotheses on how to reduce or abolish toxicity. Results We have used untargeted metabolome analysis to determine differentially expressed metabolites in human lung epithelial cells (A549) exposed to copper oxide nanoparticles (CuO NPs). Toxicity hypotheses were then generated based on the affected pathways, and critically tested using more conventional biochemical and cellular assays. CuO NPs induced regulation of metabolites involved in oxidative stress, hypertonic stress, and apoptosis. The involvement of oxidative stress was clarified more easily than apoptosis, which involved control experiments to confirm specific metabolites that could be used as standard markers for apoptosis; based on this we tentatively propose methylnicotinamide as a generic metabolic marker for apoptosis. Conclusions Our findings are well aligned with the current literature on CuO NP toxicity. We thus believe that untargeted metabolomics profiling is a suitable tool for NP toxicity screening and hypothesis generation. Keywords Untargeted metabolomics Copper oxide nanoparticles Apoptosis Oxidative stress Toxicity profiling Adverse outcome pathways}, keywords = {}, pubstate = {published}, tppubtype = {article} } Background The rapidly increasing number of engineered nanoparticles (NPs), and products containing NPs, raises concerns for human exposure and safety. With this increasing, and ever changing, catalogue of NPs it is becoming more difficult to adequately assess the toxic potential of new materials in a timely fashion. It is therefore important to develop methods which can provide high-throughput screening of biological responses. The use of omics technologies, including metabolomics, can play a vital role in this process by providing relatively fast, comprehensive, and cost-effective assessment of cellular responses. These techniques thus provide the opportunity to identify specific toxicity pathways and to generate hypotheses on how to reduce or abolish toxicity. Results We have used untargeted metabolome analysis to determine differentially expressed metabolites in human lung epithelial cells (A549) exposed to copper oxide nanoparticles (CuO NPs). Toxicity hypotheses were then generated based on the affected pathways, and critically tested using more conventional biochemical and cellular assays. CuO NPs induced regulation of metabolites involved in oxidative stress, hypertonic stress, and apoptosis. The involvement of oxidative stress was clarified more easily than apoptosis, which involved control experiments to confirm specific metabolites that could be used as standard markers for apoptosis; based on this we tentatively propose methylnicotinamide as a generic metabolic marker for apoptosis. Conclusions Our findings are well aligned with the current literature on CuO NP toxicity. We thus believe that untargeted metabolomics profiling is a suitable tool for NP toxicity screening and hypothesis generation. Keywords Untargeted metabolomics Copper oxide nanoparticles Apoptosis Oxidative stress Toxicity profiling Adverse outcome pathways |
Nicoludis, John M; Vogt, Bennett E; Green, Anna G; Schärfe, Charlotta PI; Marks, Debora S; Gaudet, Rachelle Antiparallel protocadherin homodimers use distinct affinity- and specificity-mediating regions in cadherin repeats 1-4 eLife, 5 , pp. e18449, 2016. @article{10.7554-eLife.18449, title = {Antiparallel protocadherin homodimers use distinct affinity- and specificity-mediating regions in cadherin repeats 1-4}, author = {John M Nicoludis and Bennett E Vogt and Anna G Green and Charlotta PI Schärfe and Debora S Marks and Rachelle Gaudet}, url = {https://dx.doi.org/10.7554/eLife.18449}, year = {2016}, date = {2016-01-01}, journal = {eLife}, volume = {5}, pages = {e18449}, abstract = {Protocadherins (Pcdhs) are cell adhesion and signaling proteins used by neurons to develop and maintain neuronal networks, relying on trans homophilic interactions between their extracellular cadherin (EC) repeat domains. We present the structure of the antiparallel EC1-4 homodimer of human PcdhγB3, a member of the γ subfamily of clustered Pcdhs. Structure and sequence comparisons of α, β, and γ clustered Pcdh isoforms illustrate that subfamilies encode specificity in distinct ways through diversification of loop region structure and composition in EC2 and EC3, which contains isoform-specific conservation of primarily polar residues. In contrast, the EC1/EC4 interface comprises hydrophobic interactions that provide non-selective dimerization affinity. Using sequence coevolution analysis, we found evidence for a similar antiparallel EC1-4 interaction in non-clustered Pcdh families. We thus deduce that the EC1-4 antiparallel homodimer is a general interaction strategy that evolved before the divergence of these distinct protocadherin families.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Protocadherins (Pcdhs) are cell adhesion and signaling proteins used by neurons to develop and maintain neuronal networks, relying on trans homophilic interactions between their extracellular cadherin (EC) repeat domains. We present the structure of the antiparallel EC1-4 homodimer of human PcdhγB3, a member of the γ subfamily of clustered Pcdhs. Structure and sequence comparisons of α, β, and γ clustered Pcdh isoforms illustrate that subfamilies encode specificity in distinct ways through diversification of loop region structure and composition in EC2 and EC3, which contains isoform-specific conservation of primarily polar residues. In contrast, the EC1/EC4 interface comprises hydrophobic interactions that provide non-selective dimerization affinity. Using sequence coevolution analysis, we found evidence for a similar antiparallel EC1-4 interaction in non-clustered Pcdh families. We thus deduce that the EC1-4 antiparallel homodimer is a general interaction strategy that evolved before the divergence of these distinct protocadherin families. |
Krüger, Jens; Thiel, Philipp; Merelli, Ivan; Grunzke, Richard; Gesing, Sandra Portals and Web-based Resources for Virtual Screening. Curr Drug Targets, 2016, (automatic medline import). @article{KrugerEtAl2016, title = {Portals and Web-based Resources for Virtual Screening.}, author = {Jens Krüger and Philipp Thiel and Ivan Merelli and Richard Grunzke and Sandra Gesing}, doi = {https://doi.org/10.2174/1389450117666160201105806}, year = {2016}, date = {2016-01-01}, journal = {Curr Drug Targets}, abstract = {Virtual screening for active compounds has become an essential step within the drug development pipeline. The computer based prediction of compound-binding modes is one of the most time and cost efficient methods for screening ligand libraries and enrich results of potential drugs. Here we present an overview about currently available online resources regarding compound databases, docking applications, and science gateways for drug discovery and virtual screening, in order to help structural biologists in choosing the best tools for their analysis. The appearance of the user interface, authentication and security aspects, data management, and computational performance will be discussed. We anticipate a broad overview about currently available solutions, guiding computational chemists and users from related fields towards scientifically reliable results.}, note = {automatic medline import}, keywords = {}, pubstate = {published}, tppubtype = {article} } Virtual screening for active compounds has become an essential step within the drug development pipeline. The computer based prediction of compound-binding modes is one of the most time and cost efficient methods for screening ligand libraries and enrich results of potential drugs. Here we present an overview about currently available online resources regarding compound databases, docking applications, and science gateways for drug discovery and virtual screening, in order to help structural biologists in choosing the best tools for their analysis. The appearance of the user interface, authentication and security aspects, data management, and computational performance will be discussed. We anticipate a broad overview about currently available solutions, guiding computational chemists and users from related fields towards scientifically reliable results. |
Hong, Henoch S; Koch, Sven D; Scheel, Birgit; Gnad-Vogt, Ulrike; Schröder, Andreas; Kallen, Karl-Josef; Wiegand, Volker; Backert, Linus; Kohlbacher, Oliver; Hoerr, Ingmar; Fotin-Mleczek, Mariola; Billingsley, James M Distinct transcriptional changes in non-small cell lung cancer patients associated with multi-antigenic RNActive® CV9201 immunotherapy Oncoimmunol., 5 (12), pp. e1249560., 2016. @article{OncoImmuno2016, title = {Distinct transcriptional changes in non-small cell lung cancer patients associated with multi-antigenic RNActive® CV9201 immunotherapy}, author = {Henoch S Hong and Sven D Koch and Birgit Scheel and Ulrike Gnad-Vogt and Andreas Schröder and Karl-Josef Kallen and Volker Wiegand and Linus Backert and Oliver Kohlbacher and Ingmar Hoerr and Mariola Fotin-Mleczek and James M Billingsley}, doi = {https://doi.org/10.1080/2162402X.2016.1249560}, year = {2016}, date = {2016-01-01}, journal = {Oncoimmunol.}, volume = {5}, number = {12}, pages = {e1249560.}, abstract = {We recently completed a phase I/IIa trial of RNActive® CV9201, a novel mRNA-based therapeutic vaccine targeting five tumor-associated antigens in non-small cell lung cancer (NSCLC) patients. The aim of the study presented here was to comprehensively analyze changes in peripheral blood during the vaccination period and to generate hypotheses facilitating the identification of potential biomarkers correlating with differential clinical outcomes post RNActive® immunotherapy. We performed whole-genome expression profiling in a subgroup of 22 stage IV NSCLC patients before and after initiation of treatment with CV9201. Utilizing an analytic approach based on blood transcriptional modules (BTMs), a previously described, sensitive tool for blood transcriptome data analysis, patients segregated into two major clusters based on transcriptional changes post RNActive® treatment. The first group of patients was characterized by the upregulation of an expression signature associated with myeloid cells and inflammation, whereas the other group exhibited an expression signature associated with T and NK cells. Patients with an enrichment of T and NK cell modules after treatment compared to baseline exhibited significantly longer progression-free and overall survival compared to patients with an upregulation of myeloid cell and inflammatory modules. Notably, these gene expression signatures were mutually exclusive and inversely correlated. Furthermore, our findings correlated with phenotypic data derived by flow cytometry as well as the neutrophil-to-lymphocyte ratio. Our study thus demonstrates non-overlapping, distinct transcriptional profiles correlating with survival warranting further validation for the development of biomarker candidates for mRNA-based immunotherapy.}, keywords = {}, pubstate = {published}, tppubtype = {article} } We recently completed a phase I/IIa trial of RNActive® CV9201, a novel mRNA-based therapeutic vaccine targeting five tumor-associated antigens in non-small cell lung cancer (NSCLC) patients. The aim of the study presented here was to comprehensively analyze changes in peripheral blood during the vaccination period and to generate hypotheses facilitating the identification of potential biomarkers correlating with differential clinical outcomes post RNActive® immunotherapy. We performed whole-genome expression profiling in a subgroup of 22 stage IV NSCLC patients before and after initiation of treatment with CV9201. Utilizing an analytic approach based on blood transcriptional modules (BTMs), a previously described, sensitive tool for blood transcriptome data analysis, patients segregated into two major clusters based on transcriptional changes post RNActive® treatment. The first group of patients was characterized by the upregulation of an expression signature associated with myeloid cells and inflammation, whereas the other group exhibited an expression signature associated with T and NK cells. Patients with an enrichment of T and NK cell modules after treatment compared to baseline exhibited significantly longer progression-free and overall survival compared to patients with an upregulation of myeloid cell and inflammatory modules. Notably, these gene expression signatures were mutually exclusive and inversely correlated. Furthermore, our findings correlated with phenotypic data derived by flow cytometry as well as the neutrophil-to-lymphocyte ratio. Our study thus demonstrates non-overlapping, distinct transcriptional profiles correlating with survival warranting further validation for the development of biomarker candidates for mRNA-based immunotherapy. |
Thijssen, Bram; Dijkstra, Tjeerd; Heskes, Tom; Wessels, Lodewyk BCM: toolkit for Bayesian analysis of Computational Models using samplers BMC Systems Biology, 10:100 , 2016. @article{articlereference.2016-11-08.1400143489, title = {BCM: toolkit for Bayesian analysis of Computational Models using samplers}, author = {Bram Thijssen and Tjeerd Dijkstra and Tom Heskes and Lodewyk Wessels}, url = {https://bmcsystbiol.biomedcentral.com/articles/10.1186/s12918-016-0339-3}, year = {2016}, date = {2016-01-01}, journal = {BMC Systems Biology}, volume = {10:100}, abstract = {We present BCM, a toolkit for the Bayesian analysis of Computational Models using samplers. It provides efficient, multithreaded implementations of eleven algorithms for sampling from posterior probability distributions and for calculating marginal likelihoods. BCM includes tools to simplify the process of model specification and scripts for visualizing the results. The flexible architecture allows it to be used on diverse types of biological computational models. In an example inference task using a model of the cell cycle based on ordinary differential equations, BCM is significantly more efficient than existing software packages, allowing more challenging inference problems to be solved.}, keywords = {}, pubstate = {published}, tppubtype = {article} } We present BCM, a toolkit for the Bayesian analysis of Computational Models using samplers. It provides efficient, multithreaded implementations of eleven algorithms for sampling from posterior probability distributions and for calculating marginal likelihoods. BCM includes tools to simplify the process of model specification and scripts for visualizing the results. The flexible architecture allows it to be used on diverse types of biological computational models. In an example inference task using a model of the cell cycle based on ordinary differential equations, BCM is significantly more efficient than existing software packages, allowing more challenging inference problems to be solved. |
Dietsche, Tobias; Mebrhatu, Mehari Tesfazgi; Brunner, Matthias J; Abrusci, Patrizia; Yan, Jun; Franz-Wachtel, Mirita; Schärfe, Charlotta; Zilkenat, Susann; Grin, Iwan; Galán, Jorge E; Kohlbacher, Oliver; Lea, Susan; Macek, Boris; Marlovits, Thomas C; Robinson, Carol; Wagner, Samuel Structural and functional characterization of the bacterial type III secretion export apparatus PLoS Pathogens, 12 (12), pp. e1006071., 2016. @article{PlosPath2016, title = {Structural and functional characterization of the bacterial type III secretion export apparatus}, author = {Tobias Dietsche and Mehari Tesfazgi Mebrhatu and Matthias J Brunner and Patrizia Abrusci and Jun Yan and Mirita Franz-Wachtel and Charlotta Schärfe and Susann Zilkenat and Iwan Grin and Jorge E Galán and Oliver Kohlbacher and Susan Lea and Boris Macek and Thomas C Marlovits and Carol Robinson and Samuel Wagner}, doi = {https://doi.org/10.1371/journal.ppat.1006071}, year = {2016}, date = {2016-01-01}, journal = {PLoS Pathogens}, volume = {12}, number = {12}, pages = {e1006071.}, abstract = {Bacterial type III protein secretion systems inject effector proteins into eukaryotic host cells in order to promote survival and colonization of Gram-negative pathogens and symbionts. Secretion across the bacterial cell envelope and injection into host cells is facilitated by a so-called injectisome. Its small hydrophobic export apparatus components SpaP and SpaR were shown to nucleate assembly of the needle complex and to form the central “cup” substructure of a Salmonella Typhimurium secretion system. However, the in vivo placement of these components in the needle complex and their function during the secretion process remained poorly defined. Here we present evidence that a SpaP pentamer forms a 15 Å wide pore and provide a detailed map of SpaP interactions with the export apparatus components SpaQ, SpaR, and SpaS. We further refine the current view of export apparatus assembly, consolidate transmembrane topology models for SpaP and SpaR, and present intimate interactions of the periplasmic domains of SpaP and SpaR with the inner rod protein PrgJ, indicating how export apparatus and needle filament are connected to create a continuous conduit for substrate translocation.}, keywords = {}, pubstate = {published}, tppubtype = {article} } Bacterial type III protein secretion systems inject effector proteins into eukaryotic host cells in order to promote survival and colonization of Gram-negative pathogens and symbionts. Secretion across the bacterial cell envelope and injection into host cells is facilitated by a so-called injectisome. Its small hydrophobic export apparatus components SpaP and SpaR were shown to nucleate assembly of the needle complex and to form the central “cup” substructure of a Salmonella Typhimurium secretion system. However, the in vivo placement of these components in the needle complex and their function during the secretion process remained poorly defined. Here we present evidence that a SpaP pentamer forms a 15 Å wide pore and provide a detailed map of SpaP interactions with the export apparatus components SpaQ, SpaR, and SpaS. We further refine the current view of export apparatus assembly, consolidate transmembrane topology models for SpaP and SpaR, and present intimate interactions of the periplasmic domains of SpaP and SpaR with the inner rod protein PrgJ, indicating how export apparatus and needle filament are connected to create a continuous conduit for substrate translocation. |
Krüger, Jens; Kohlbacher, Oliver Containerization and wrapping of a mass spectra prediction workflow 2016. @preprint{KruegerKohlbacher2016, title = {Containerization and wrapping of a mass spectra prediction workflow}, author = {Jens Krüger and Oliver Kohlbacher}, url = {https://peerj.com/preprints/2528/}, year = {2016}, date = {2016-01-01}, abstract = {Practical experiences are reported about implementing a workflow for the prediction of mass spectra. QCEIMS is used to simulate the fragmentation trajectories consequently leading to predicted mass spectra for small molecules, such as metabolites. The individual calculations are embedded into UNICORE workflow nodes using Docker containerization for the applications themselves. Challenges, caveats, but also advantages are discussed, providing guidance for the deployment of a scientific protocol on high performance computing resources.}, keywords = {}, pubstate = {published}, tppubtype = {preprint} } Practical experiences are reported about implementing a workflow for the prediction of mass spectra. QCEIMS is used to simulate the fragmentation trajectories consequently leading to predicted mass spectra for small molecules, such as metabolites. The individual calculations are embedded into UNICORE workflow nodes using Docker containerization for the applications themselves. Challenges, caveats, but also advantages are discussed, providing guidance for the deployment of a scientific protocol on high performance computing resources. |
Aebersold, Rudolf; Kohlbacher, Oliver; Vitek, Olga Computational Mass Spectrometry (Dagstuhl Seminar 15351) Dagstuhl Reports, 5 (8), pp. 9–33, 2016, ISSN: 2192-5283. @article{aebersold_et_al:DR:2016:5676, title = {Computational Mass Spectrometry (Dagstuhl Seminar 15351)}, author = {Rudolf Aebersold and Oliver Kohlbacher and Olga Vitek}, editor = {Rudolf Aebersold and Oliver Kohlbacher and Olga Vitek}, url = {http://drops.dagstuhl.de/opus/volltexte/2016/5676}, doi = {10.4230/DagRep.5.8.9}, issn = {2192-5283}, year = {2016}, date = {2016-01-01}, journal = {Dagstuhl Reports}, volume = {5}, number = {8}, pages = {9--33}, publisher = {Schloss Dagstuhl--Leibniz-Zentrum fuer Informatik}, address = {Dagstuhl, Germany}, keywords = {}, pubstate = {published}, tppubtype = {article} } |
Hildebrandt, Anna Katharina; Stöckel, Daniel; Fischer, Nina; de la Trevino, Luis Garza; Krüger, Jens; Nickels, Stefan; Röttig, Marc; Schärfe, Charlotta; Schumann, Marcel; Thiel, Philipp; Lenhof, Hans-Peter; Kohlbacher, Oliver; Hildebrandt, Andreas ballaxy: web services for structural bioinformatics Bioinformatics, 31 (1), pp. 121-2, 2015. @article{BALLaxyBioinfo, title = {ballaxy: web services for structural bioinformatics}, author = {Anna Katharina Hildebrandt and Daniel Stöckel and Nina Fischer and Luis de la Garza Trevino and Jens Krüger and Stefan Nickels and Marc Röttig and Charlotta Schärfe and Marcel Schumann and Philipp Thiel and Hans-Peter Lenhof and Oliver Kohlbacher and Andreas Hildebrandt}, doi = {https://doi.org/10.1093/bioinformatics/btu574}, year = {2015}, date = {2015-01-01}, journal = {Bioinformatics}, volume = {31}, number = {1}, pages = {121-2}, abstract = {MOTIVATION: Web-based workflow systems have gained considerable momentum in sequence-oriented bioinformatics. In structural bioinformatics, however, such systems are still relatively rare: while commercial stand-alone workflow applications are common in the pharmaceutical industry, academic researchers often still rely on command-line scripting to glue individual tools together. RESULTS: In this work, we address the problem of building a web-based system for workflows in structural bioinformatics. For the underlying molecular modelling engine, we opted for the BALL framework (Hildebrandt et al., 2010) due to its extensive and well tested functionality in the field of structural bioinformatics. The large number of molecular data structures and algorithms implemented in BALL allows for elegant and sophisticated development of new approaches in the field. We hence connected the versatile BALL library and its visualization and editing frontend BALLView with the Galaxy workflow framework. The result, which we call ballaxy, enables the user to simply and intuitively create sophisticated pipelines for applications in structure-based computational biology, integrated into a standard tool for molecular modelling. AVAILABILITY: ballaxy consists of three parts: some minor modifications to the Galaxy system, a collection of tools, and an integration into the BALL-framework and the BALLView application for molecular modelling. Modifications to Galaxy will be submitted to the Galaxy project, and the BALL and BALLView integrations will be integrated in the next major BALL release. After acceptance of the modifications into the Galaxy project, we will publish all ballaxy tools via the Galaxy toolshed. In the meantime, all three components are available from http://www.ball-project.org/ballaxy. Ballaxy is licensed under the terms of the GPL.}, keywords = {}, pubstate = {published}, tppubtype = {article} } MOTIVATION: Web-based workflow systems have gained considerable momentum in sequence-oriented bioinformatics. In structural bioinformatics, however, such systems are still relatively rare: while commercial stand-alone workflow applications are common in the pharmaceutical industry, academic researchers often still rely on command-line scripting to glue individual tools together. RESULTS: In this work, we address the problem of building a web-based system for workflows in structural bioinformatics. For the underlying molecular modelling engine, we opted for the BALL framework (Hildebrandt et al., 2010) due to its extensive and well tested functionality in the field of structural bioinformatics. The large number of molecular data structures and algorithms implemented in BALL allows for elegant and sophisticated development of new approaches in the field. We hence connected the versatile BALL library and its visualization and editing frontend BALLView with the Galaxy workflow framework. The result, which we call ballaxy, enables the user to simply and intuitively create sophisticated pipelines for applications in structure-based computational biology, integrated into a standard tool for molecular modelling. AVAILABILITY: ballaxy consists of three parts: some minor modifications to the Galaxy system, a collection of tools, and an integration into the BALL-framework and the BALLView application for molecular modelling. Modifications to Galaxy will be submitted to the Galaxy project, and the BALL and BALLView integrations will be integrated in the next major BALL release. After acceptance of the modifications into the Galaxy project, we will publish all ballaxy tools via the Galaxy toolshed. In the meantime, all three components are available from http://www.ball-project.org/ballaxy. Ballaxy is licensed under the terms of the GPL. |