U.S. flag

An official website of the United States government

The .gov means it’s official. Federal government websites often end in .gov or .mil. Before sharing sensitive information, make sure you’re on a federal government site.

The site is secure. The https:// ensures that you are connecting to the official website and that any information you provide is encrypted and transmitted securely.

  • Publications
  • Account settings
  • My Bibliography
  • Collections
  • Citation manager

Save citation to file

Email citation, add to collections.

  • Create a new collection
  • Add to an existing collection

Add to My Bibliography

Your saved search, create a file for external citation management software, your rss feed.

  • Search in PubMed
  • Search in NLM Catalog
  • Add to Search

Systems biology primer: the basic methods and approaches

Affiliations.

  • 1 Department of Pharmacological Sciences and Systems Biology Center New York, Icahn School of Medicine at Mount Sinai, New York, NY 10029, U.S.A.
  • 2 Department of Pharmacological Sciences and Systems Biology Center New York, Icahn School of Medicine at Mount Sinai, New York, NY 10029, U.S.A. [email protected].
  • PMID: 30287586
  • DOI: 10.1042/EBC20180003

Systems biology is an integrative discipline connecting the molecular components within a single biological scale and also among different scales (e.g. cells, tissues and organ systems) to physiological functions and organismal phenotypes through quantitative reasoning, computational models and high-throughput experimental technologies. Systems biology uses a wide range of quantitative experimental and computational methodologies to decode information flow from genes, proteins and other subcellular components of signaling, regulatory and functional pathways to control cell, tissue, organ and organismal level functions. The computational methods used in systems biology provide systems-level insights to understand interactions and dynamics at various scales, within cells, tissues, organs and organisms. In recent years, the systems biology framework has enabled research in quantitative and systems pharmacology and precision medicine for complex diseases. Here, we present a brief overview of current experimental and computational methods used in systems biology.

Keywords: Systems Biology; bioinformatics; biological networks; computational biology; computational models; personalized medicine.

© 2018 The Author(s). Published by Portland Press Limited on behalf of the Biochemical Society.

PubMed Disclaimer

Similar articles

  • Boolean modeling in systems biology: an overview of methodology and applications. Wang RS, Saadatpour A, Albert R. Wang RS, et al. Phys Biol. 2012 Oct;9(5):055001. doi: 10.1088/1478-3975/9/5/055001. Epub 2012 Sep 25. Phys Biol. 2012. PMID: 23011283 Review.
  • Deciphering Combinatorial Genetics. Wong AS, Choi GC, Lu TK. Wong AS, et al. Annu Rev Genet. 2016 Nov 23;50:515-538. doi: 10.1146/annurev-genet-120215-034902. Epub 2016 Oct 6. Annu Rev Genet. 2016. PMID: 27732793 Review.
  • Systems medicine: evolution of systems biology from bench to bedside. Wang RS, Maron BA, Loscalzo J. Wang RS, et al. Wiley Interdiscip Rev Syst Biol Med. 2015 Jul-Aug;7(4):141-61. doi: 10.1002/wsbm.1297. Epub 2015 Apr 17. Wiley Interdiscip Rev Syst Biol Med. 2015. PMID: 25891169 Free PMC article. Review.
  • Systems biology by the rules: hybrid intelligent systems for pathway modeling and discovery. Bosl WJ. Bosl WJ. BMC Syst Biol. 2007 Feb 15;1:13. doi: 10.1186/1752-0509-1-13. BMC Syst Biol. 2007. PMID: 17408503 Free PMC article.
  • Introduction: Cancer Gene Networks. Clarke R. Clarke R. Methods Mol Biol. 2017;1513:1-9. doi: 10.1007/978-1-4939-6539-7_1. Methods Mol Biol. 2017. PMID: 27807826
  • COVID-19: Recent Insight in Genomic Feature, Pathogenesis, Immunological Biomarkers, Treatment Options and Clinical Updates on SARS-CoV-2. Deshmukh R, Harwansh RK, Garg A, Mishra S, Agrawal R, Jangde R. Deshmukh R, et al. Curr Genomics. 2024 Apr 8;25(2):69-87. doi: 10.2174/0113892029291098240129113500. Curr Genomics. 2024. PMID: 38751601 Review.
  • Identification of metabolic pathways and key genes associated with atypical parkinsonism using a systems biology approach. Pasqualotto A, da Silva V, Pellenz FM, Schuh AFS, Schwartz IVD, Siebert M. Pasqualotto A, et al. Metab Brain Dis. 2024 Apr;39(4):577-587. doi: 10.1007/s11011-024-01342-7. Epub 2024 Feb 2. Metab Brain Dis. 2024. PMID: 38305999
  • How Can Proteomics Help to Elucidate the Pathophysiological Crosstalk in Muscular Dystrophy and Associated Multi-System Dysfunction? Dowling P, Trollet C, Negroni E, Swandulla D, Ohlendieck K. Dowling P, et al. Proteomes. 2024 Jan 16;12(1):4. doi: 10.3390/proteomes12010004. Proteomes. 2024. PMID: 38250815 Free PMC article.
  • Mycobacterium smegmatis , a Promising Vaccine Vector for Preventing TB and Other Diseases: Vaccinomics Insights and Applications. Xie W, Wang L, Luo D, Soni V, Rosenn EH, Wang Z. Xie W, et al. Vaccines (Basel). 2023 Jul 31;11(8):1302. doi: 10.3390/vaccines11081302. Vaccines (Basel). 2023. PMID: 37631870 Free PMC article. Review.
  • Sex-Gender-Based Differences in Metabolic Diseases. Campesi I, Ruoppolo M, Franconi F, Caterino M, Costanzo M. Campesi I, et al. Handb Exp Pharmacol. 2023;282:241-257. doi: 10.1007/164_2023_683. Handb Exp Pharmacol. 2023. PMID: 37528324

Publication types

  • Search in MeSH

Related information

  • Cited in Books

LinkOut - more resources

Full text sources.

  • Silverchair Information Systems

full text provider logo

  • Citation Manager

NCBI Literature Resources

MeSH PMC Bookshelf Disclaimer

The PubMed wordmark and PubMed logo are registered trademarks of the U.S. Department of Health and Human Services (HHS). Unauthorized use of these marks is strictly prohibited.

Discover the Top 75 Free Courses for August

experimental methods in systems biology

Udemy Announces Layoffs Without Saying ‘Layoffs’

Udemy’s latest ‘Strategic Business Update’ uses corporate euphemisms to signal job cuts while pivoting to enterprise clients.

  • 7 Best Climate Change Courses for 2024: Exploring the Science
  • [2024] 110+ Hours of Free LinkedIn Learning Courses with Free Certification
  • 7 Best Free Haskell Courses for 2024
  • 11 Best Data Structures & Algorithms Courses for 2024
  • 9 Best TensorFlow Courses for 2024

600 Free Google Certifications

Most common

  • web development

Popular subjects

Communication Skills

Computer Science

Graphic Design

Popular courses

Paleontology: Theropod Dinosaurs and the Origin of Birds

Competitive Strategy

Comprendere la filosofia

Organize and share your learning with Class Central Lists.

View our Lists Showcase

Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

Experimental Methods in Systems Biology

Icahn School of Medicine at Mount Sinai via Coursera Help

  • Introduction
  • Description goes here
  • Deep mRNA Sequencing
  • Mass Spectrometry-Based Proteomics
  • Midterm Exam
  • Flow and Mass Cytometry for Single Cell Protein Levels and Cell Fate
  • Live-cell Imaging for Single Cell Protein Dynamics
  • Integrating and Interpreting Datasets with Network Models and Dynamical Models

Marc Birtwistle

  • united states

Related Courses

Dynamical modeling methods for systems biology, introduction to systems biology, related articles, 1700 coursera courses that are still completely free, 250 top free coursera courses of all time, massive list of mooc-based microcredentials.

4.5 rating at Coursera based on 310 ratings

Select rating

Start your review of Experimental Methods in Systems Biology

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Thank you for visiting nature.com. You are using a browser version with limited support for CSS. To obtain the best experience, we recommend you use a more up to date browser (or turn off compatibility mode in Internet Explorer). In the meantime, to ensure continued support, we are displaying the site without styles and JavaScript.

  • View all journals
  • Explore content
  • About the journal
  • Publish with us
  • Sign up for alerts
  • Research Highlights
  • Published: December 2006

Systems biology for beginners

  • Daniel Evanko  

Nature Methods volume  3 ,  pages 964–965 ( 2006 ) Cite this article

588 Accesses

3 Citations

Metrics details

A web and print Focus on systems biology from Nature Publishing Group provides a practical introduction to a field that for all its promise still has many skeptics.

This is a preview of subscription content, access via your institution

Access options

Subscribe to this journal

Receive 12 print issues and online access

251,40 € per year

only 20,95 € per issue

Buy this article

  • Purchase on Springer Link
  • Instant access to full article PDF

Prices may be subject to local taxes which are calculated during checkout

You can also search for this author in PubMed   Google Scholar

Related links

The Systems biology: a user's guide Focus homepage

Rights and permissions

Reprints and permissions

About this article

Cite this article.

Evanko, D. Systems biology for beginners. Nat Methods 3 , 964–965 (2006). https://doi.org/10.1038/nmeth1206-964b

Download citation

Issue Date : December 2006

DOI : https://doi.org/10.1038/nmeth1206-964b

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

Quick links

  • Explore articles by subject
  • Guide to authors
  • Editorial policies

Sign up for the Nature Briefing newsletter — what matters in science, free to your inbox daily.

experimental methods in systems biology

Manage cookies

In this window you can switch the placement of cookies on and off. Only the Functional Cookies cannot be switched off. Your cookies preferences will be stored for 9 months, after which they will be set to default again. You can always manage your preferences and revoke your marketing consent via the Cookie Statement page in the lower left corner of the website.

Like you, we love to learn. That is why we and third parties we work with, use functional and analytical cookies (and comparable techniques) to improve your experience. By accepting all cookies you allow us to place cookies for marketing purposes too, in line with our Privacy Policy . These allow us to track your surfing behavior on and outside of Springest so we and third parties can adapt our website, advertisements and communication to you. Find more information and manage your preferences via the Cookie statement .

check Accept all Cookies

cancel Reject optional Cookies

settings Manage cookies

Experimental Methods in Systems Biology

experimental methods in systems biology

  • Price completeness: This price is complete, there are no hidden additional costs.

Logo Coursera (CC)

Need more information? Get more details on the site of the provider.

Description

When you enroll for courses through Coursera you get to choose for a paid plan or for a free plan . 

  • Free plan: No certicification and/or audit only. You will have access to all course materials except graded items.
  • Paid plan:  Commit to earning a Certificate—it's a trusted, shareable way to showcase your new skills.

About this course: Learn about the technologies underlying experimentation used in systems biology, with particular focus on RNA sequencing, mass spec-based proteomics, flow/mass cytometry and live-cell imaging. A key driver of the systems biology field is the technology allowing us to delve deeper and wider into how cells respond to experimental perturbations. This in turns allows us to build more detailed quantitative models of cellular function, which can give important insight into applications ranging from biotechnology to human disease. This course gives a broad overview of a variety of current experimental techniques used in modern systems biology, with focus on obtaining the qua…

Read the complete description

Frequently asked questions

There are no frequently asked questions yet. If you have any more questions or need help, contact our customer service .

Didn't find what you were looking for? See also: Biology , Health Management , Anatomy , Palliative Care , and Science .

About this course: Learn about the technologies underlying experimentation used in systems biology, with particular focus on RNA sequencing, mass spec-based proteomics, flow/mass cytometry and live-cell imaging. A key driver of the systems biology field is the technology allowing us to delve deeper and wider into how cells respond to experimental perturbations. This in turns allows us to build more detailed quantitative models of cellular function, which can give important insight into applications ranging from biotechnology to human disease. This course gives a broad overview of a variety of current experimental techniques used in modern systems biology, with focus on obtaining the quantitative data needed for computational modeling purposes in downstream analyses. We dive deeply into four technologies in particular, mRNA sequencing, mass spectrometry-based proteomics, flow/mass cytometry, and live-cell imaging. These techniques are often used in systems biology and range from genome-wide coverage to single molecule coverage, millions of cells to single cells, and single time points to frequently sampled time courses. We present not only the theoretical background upon which these technologies work, but also enter real wet lab environments to provide instruction on how these techniques are performed in practice, and how resultant data are analyzed for quality and content.

Taught by:   Marc Birtwistle, PhD, Assistant Professor

Cada curso es como un libro de texto interactivo, con videos pregrabados, cuestionarios y proyectos.

Conéctate con miles de estudiantes y debate ideas y materiales del curso, y obtén ayuda para dominar los conceptos.

Obtén reconocimiento oficial por tu trabajo y comparte tu éxito con amigos, compañeros y empleadores.

  • Video: Lecture 1 -Scope and Overview
  • Leyendo: Lecture Slides
  • Video: Lecture 2 - Biological Model Systems
  • Video: Lecture 3 - Experimental Perturbations
  • Leyendo: Lecture Slide
  • Video: Lecture 4 - Measuring Nucleic Acids
  • Video: Lecture 5 - Measuring Protein and Protein States
  • Leyendo: Reading List
  • Video: Lecture 1 - History of Sequencing
  • Video: Lecture 2 - 2nd and 3rd Generation Sequencing
  • Video: Lecture 3 - Illumina-Based mRNA Sequencing
  • Video: Lecture 4 - Lab-Based Video
  • Video: Lecture 5 - mRNA Sequencing Data Analysis
  • Video: Lecture 1 - Basics of Mass Spectrometry 1
  • Video: Lecture 2 - Basics of Mass Spectrometry 2
  • Video: Lecture 3 - Quantification in Proteomics
  • Video: Lecture 4 - Lab-Based Video - Proteomics Day 1
  • Video: Lecture 5 - Lab-Based Video - Proteomics Day 2
  • Video: Lecture 6 - Lab-Based Video - Analysis
  • Video: Lecture 1 - Flow Cytometry 1
  • Video: Lecture 2 - Flow Cytometry 2
  • Video: Lecture 3 - Mass Cytometry
  • Video: Lecture 4 - Cytometry Data Analysis
  • Video: Lecture 5 - Lab-Based Video - Flow Cytometry - Acquisition
  • Video: Lecture 6 - Lab-Based Video - Flow Cytometry - Analysis
  • Video: Lecture 7 - Lab-Based Video - Mass Cytometry
  • Video: Lecture 1 - Fluorescence Microscopy
  • Video: Lecture 2 - Types of Imaging
  • Video: Lecture 3 - Visualizing Molecules in Living Cells: Fluorescent Tools
  • Video: Lecture 4 - Quantification
  • Video: Lecture 5 - Lab-Based Video
  • Video: Lecture 1 - Omics data and Network Model Analyses
  • Video: Lecture 2 - Single Cell Time Course Data and Dynamical Model Analyses
  • Video: Lecture 3 - Dynamical Model Case Study
  • View related products with reviews: Biology .

Share your review

  • Privacy Policy
  • Cookie statement
  • Your reference ID is: X8L9F
  • The Netherlands
  • United Kingdom

experimental methods in systems biology

Experimental Methods in Systems Biology

experimental methods in systems biology

  • From www.coursera.org
  • Free Access
  • Fee-based Certificate
  • 8 Sequences
  • Introductive Level

experimental methods in systems biology

Their employees are learning daily with

Course details.

  • Week 1 - Introduction Description goes here
  • Week 2 - Deep mRNA Sequencing Description goes here
  • Week 3 - Mass Spectrometry-Based Proteomics Description goes here
  • Week 4 - Midterm Exam Description goes here
  • Week 5 - Flow and Mass Cytometry for Single Cell Protein Levels and Cell Fate Description goes here
  • Week 6 - Live-cell Imaging for Single Cell Protein Dynamics Description goes here
  • Week 7 - Integrating and Interpreting Datasets with Network Models and Dynamical Models Description goes here
  • Week 8 - Final Exam  

Prerequisite

Instructors.

Marc Birtwistle, PhD Assistant Professor Department of Pharmacology and Systems Therapeutics, Systems Biology Center New York (SBCNY)

Coursera is a digital company offering massive open online course founded by computer teachers Andrew Ng and Daphne Koller Stanford University, located in Mountain View, California. 

Coursera works with top universities and organizations to make some of their courses available online, and offers courses in many subjects, including: physics, engineering, humanities, medicine, biology, social sciences, mathematics, business, computer science, digital marketing, data science, and other subjects.

experimental methods in systems biology

Get certified in

10,000+ Free Udemy Courses to Start Today

Close icon

  • Free Courses
  • Universities
  • Best Courses
  • Submit Course

Coursesity is supported by learner community. We may earn affiliate commission when you make purchase via links on Coursesity.

Experimental Methods in Systems Biology

Experimental Methods in Systems Biology

Learn experimental methods in systems biology from icahn school of medicine at mount sinai..

experimental methods in systems biology

total enrollments

experimental methods in systems biology

Total ratings

  • Course Overview

Description

In this course, you will :

  • Learn about the technologies underlying experimentation used in systems biology, with particular focus on RNA sequencing, mass spec-based proteomics, flow/mass cytometry and live-cell imaging.
  • A key driver of the systems biology field is the technology allowing us to delve deeper and wider into how cells respond to experimental perturbations. 
  • This course gives a broad overview of a variety of current experimental techniques used in modern systems biology, with focus on obtaining the quantitative data needed for computational modeling purposes in downstream analyses. 

Similar Courses

experimental methods in systems biology

Women in environmental biology

experimental methods in systems biology

Plant Cell Bioprocessing

experimental methods in systems biology

Introduction to Biology: Ecology, Evolution, & Biodiversity

experimental methods in systems biology

Essential Human Biology: Cells and Tissues

experimental methods in systems biology

Biology101: Nervous, Endocrine and Skeletal System

experimental methods in systems biology

Science Matters: Let's Talk About COVID-19

experimental methods in systems biology

Biology 101: Introduction to Digestion and Excretion

experimental methods in systems biology

Free Biology Tutorial - A-Level Biology - An introduction to key concepts.

experimental methods in systems biology

Introductory Biology

Course features.

  • Certificate on purchase
  • Coursera Icahn School of Medicine at Mount Sinai
  • Intermediate
  • Science ,Biology

Enrollment options

  • Course Material
  • Graded Assignment
  • Practice Quizzes
  • Certificate on completion
  • 7 - days Free access
  • $49/month (after free trial)
  • Cancel subscription anytime

Coursera Plus - Monthly

  • 7 - days free access
  • $59/month (after trial ends)
  • Unlimited Access to 3000+ courses

Coursera Plus - Annual

  • $399/year (43% saving)
  • 14 - days refund

Looking for the financial Help? Apply

  • Methodology Article
  • Open access
  • Published: 01 December 2014

A Bayesian active learning strategy for sequential experimental design in systems biology

  • Edouard Pauwels 1 , 2 ,
  • Christian Lajaunie 3 , 4 , 5 &
  • Jean-Philippe Vert 3 , 4 , 5  

BMC Systems Biology volume  8 , Article number:  102 ( 2014 ) Cite this article

4104 Accesses

16 Citations

Metrics details

Dynamical models used in systems biology involve unknown kinetic parameters. Setting these parameters is a bottleneck in many modeling projects. This motivates the estimation of these parameters from empirical data. However, this estimation problem has its own difficulties, the most important one being strong ill-conditionedness. In this context, optimizing experiments to be conducted in order to better estimate a system’s parameters provides a promising direction to alleviate the difficulty of the task.

Borrowing ideas from Bayesian experimental design and active learning, we propose a new strategy for optimal experimental design in the context of kinetic parameter estimation in systems biology. We describe algorithmic choices that allow to implement this method in a computationally tractable way and make it fully automatic. Based on simulation, we show that it outperforms alternative baseline strategies, and demonstrate the benefit to consider multiple posterior modes of the likelihood landscape, as opposed to traditional schemes based on local and Gaussian approximations.

This analysis demonstrates that our new, fully automatic Bayesian optimal experimental design strategy has the potential to support the design of experiments for kinetic parameter estimation in systems biology.

Systems biology emerged a decade ago as the study of biological systems where interactions between relatively simple biological species generate overall complex phenomena [ 1 ]. Quantitative mathematical models, coupled with experimental work, now play a central role to analyze, simulate and predict the behavior of biological systems. For example, ordinary differential equation- (ODE) based models, which are the focus of this work, have proved very useful to model numerous regulatory, signaling and metabolic pathways [ 2 ]-[ 4 ], including for example the cell cycle in budding yeast [ 5 ], the regulatory module of nuclear factor θ B (NF- θ B) signaling pathway [ 6 ],[ 7 ], the MAP kinase signaling pathways [ 8 ] or the caspase function in apoptosis [ 9 ].

Such dynamical models involve unknown parameters, such as kinetic parameters, that one must guess from prior knowledge or estimate from experimental data in order to analyze and simulate the model. Setting these parameters is often challenging, and constitutes a bottleneck in many modeling project [ 3 ],[ 10 ]. On the one hand, fixing parameters from estimates obtained in vitro with purified proteins may not adequately reflect the true activity in the cell, and is usually only feasible for a handful of parameters. On the other hand, optimizing parameters to reflect experimental data on how some observables behave under various experimental conditions is also challenging, since some parameters may not be identifiable, or may only be estimated with a large errors, due to the frequent lack of systematic quantitative measurements covering all variables involved in the system; many authors found, for example, that finding parameters to fit experimental observations in nonlinear models is a very ill-conditioned and multimodal problem, a phenomenon sometimes referred to as sloppiness [ 11 ]-[ 17 ], a concept closely related to that of identifiability in system identification theory [ 18 ],[ 19 ], see also [ 20 ] for a recent review. When the system has more than a few unknown parameters, computational issues also arise to efficiently sample the space of parameters [ 21 ],[ 22 ], which has been found to be very rugged and sometimes misleading in the sense that many sets of parameters that have a good fit to experimental data are meaningless from a biological point of view [ 23 ].

Optimizing the experiments to be conducted in order to alleviate non-identifiabilities and better estimate a system’s parameters therefore provides a promising direction to alleviate the difficulty of the task, and has already been the subject of much research in systems biology [ 20 ],[ 24 ]. Some authors have proposed strategies involving random sampling of parameters near the optimal one, or at least coherent with available experimental observations, and systematic simulations of the model with these parameters in order to identify experiments that would best reduce the uncertainty about the parameters [ 25 ]-[ 27 ]. A popular way to formalize and implement this idea is to follow the theory of Bayesian optimal experimental design (OED) [ 28 ],[ 29 ]. In this framework, approximating the model by a linear model (and the posterior distribution by a normal distribution) leads to the well-known A-optimal [ 30 ],[ 31 ] or D-optimal [ 32 ]-[ 36 ] experimental designs, which optimize a property of the Fisher information matrix (FIM) at the maximum likelihood estimator. FIM-based methods have the advantage to be simple and computationally efficient, but the drawback is that the assumption that the posterior probability is well approximated by a unimodal, normal distribution is usually too strong. To overcome this difficulty at the expense of computational burden, other methods involving a sampling of the posterior distribution by Monte-Carlo Markov chain (MCMC) techniques have also been proposed [ 37 ],[ 38 ]. When the goal of the modeling approach is not to estimate the parameters per se , but to understand and simulate the system, other authors have also considered the problem of experimental design to improve the predictions made by the model [ 39 ]-[ 41 ], or to discriminate between different candidate models [ 42 ]-[ 45 ].

In this work we propose a new general strategy for Bayesian OED, and study its relevance for kinetic parameter estimation in the context of systems biology. As opposed to classical Bayesian OED strategies which select the experiment that most reduces the uncertainty in parameter estimation, itself quantified by the variance or the entropy of the posterior parameter distribution, we formulate the problem in a decision-theoretic framework where we wish to minimize an error function quantifying how far the estimated parameters are from the true ones. For example, if we focus on the squared error between the estimated and true parameters, our methods attempts to minimize not only the variance of the estimates, as in standard A-optimal designs [ 30 ],[ 31 ], but also a term related to the bias of the estimate. This idea is similar to an approach that was proposed for active learning [ 46 ], where instead of just reducing the size of the version space (i.e., the amount of models coherent with observed data) the authors propose to directly optimize a loss function relevant for the task at hand. Since the true parameter needed to define the error function is unknown, we follow an approach similar to [ 46 ] and average the error function according to the current prior on the parameters. This results in a unique, well-defined criterion that can be evaluated and used to select an optimal experiment.

In the rest of this paper, we provide a rigorous derivation of this criterion, and discuss different computational strategies to evaluate it efficiently. The criterion involves an average over the parameter space according to a prior distribution, for wich we designed an exploration strategy that proved to be efficient in our experiments. We implemented the criterion in the context of an iterative experimental design problem, where a succession of experiments with different costs is allowed and the goal is to reach the best final parameter estimation given a budget to be spent, a problem that was made popular by the DREAM 6 and DREAM 7 Network Topology and Parameter Inference Challenge [ 47 ]-[ 49 ]. We demonstrate the relevance of our new OED strategy on a small simulated network in this context, and illustrate its behavior on the DREAM7 challenge. The method is fully automated, and we provide an R R package to reproduce all simulations.

A new criterion for Bayesian OED

In this section we propose a new, general criterion for Bayesian OED. We consider a system whose behavior and observables are controlled by an unknown parameter θ * ∈ ⊂ ℝ p that we wish to estimate. For that purpose, we can design an experiment e 2, which in our application will include which observables we observe, when, and under which experimental conditions. The completion of the experiment will lead to an observation o , which we model as a random variable generated according to the distribution o ~ P ( o | θ * ; e ). Note that although θ * is unknown, the distribution P ( o | θ ; e ) is supposed to be known for any θ and e , and amenable to simulations; in our case, P ( o | θ ; e ) typically involves the dynamical equations of the system if the parameters are known, and the noise model of the observations.

Our goal is to propose a numerical criterion to quantify how "good" the choice of the experiment e is for the purpose of evaluating θ θ . For that purpose, we assume given a loss function θ such that ` ( θ , θ * ) measures the loss associated to an estimate θ when the true value is θ * . A typical loss function is the squared Euclidean distance l ( θ , θ * )=║ θ - θ * ║ 2 , or the squared Euclidean distance in after a log transform for positive parameters l ( θ , θ * ) = ∑ i = 1 p log ( θ i / θ i * ) 2 . We place ourselves in a Bayesian setting, where instead of a single point estimate the knowledge about θ * at a given stage of the analysis is represented by a probability distribution π over θ . The quality of the information it provides can be quantified by the average loss, or risk:

Once we choose an experiment e and observe o , the knowledge about θ * is updated and encoded in the posterior distribution

whose risk is now:

The above expression is for a particular observation o . This observation is actually generated according to P ( o | θ * ; e ). Accordingly, the average risk of the experiment e (if the true parameter is θ * ) is:

Finally, θ * being unknown, we average the risk by taking account of the current state of knowledge, and thus according to π . The expected risk associated to the choice of e when the current knowledge about θ * is encoded in the distribution π is thus:

The expected risk R ( e ; π ) of a candidate experiment e given our current estimate of the parameter distribution π is the criterion we propose in order to assess the relevance of performing e . In other words, given a current estimate π , we propose to select the best experiment to perform as the one that minimizes R ( e ; π ). We describe in the next section more precisely how to use this criterion in the context of sequential experimental design where each experiment has a cost.

Note that the criterion R ( e ; π ) is similar but different from classical Bayesian OED criteria, like the variance criterion used in A-optimal design. Indeed, taking for example the square Euclidean loss as loss function l ( θ , θ * )=║ θ - θ * ║ 2 , and denoting by π e the mean posterior distribution that we expect if we perform experiment e , standard A-optimal design tries to minimize the variance of π e , while our criterion focuses on:

In other words, our criterion attempts to control both the bias and the variance of the posterior distribution, while standard Bayesian OED strategies only focus on the variance terms. While both criteria coincide with unbiased estimators, there is often no reason to believe that the estimates used are unbiased.

Sequential experimental design

In sequential experimental design, we sequentially choose an experiment to perform, and observe the resulting outcome. Given the past experiments e 1 ,. . ., e k and corresponding observations o 1 ,. . ., o k , we therefore need to choose what is the best next experiment e k +1 to perform, assuming in addition that each possible experiment e has an associated cost C e and we have a limited total budget to spend.

We denote by π k the distribution on π representing our knowledge about θ * after the k -th experiment and observation, with π 0 representing the prior knowledge we may have about the parameters before the first experiment. According to Bayes’ rule ( 1 ), the successive posteriors are related to each other according to:

Although a global optimization problem could be written to optimize the choice of the k -th experiment based on possible future observations and total budget constraint, we propose a simple, greedy formulation where at each step we choose the experiment that most decreases the estimation risk per cost unit. If the cost of all experiments were the same, this would simply translate to:

To take into account the different costs associated with different experiments, we consider as a baseline the mean risk when the knowledge about θ * is encoded in a distribution π over θ :

and choose the experiment that maximally reduces the expected risk per cost unit according to:

Evaluating the risk

The expected risk of an experiment R ( e ; π ) ( 2 ) involves a double integral over the parameter space and an integral over the possible observations, a challenging setting for practical evaluation. Since no analytical formula can usually be derived to compute it exactly, we now present a numerical scheme that we found efficient in practice. Since the distribution π k over the parameter space after the k -th experiment can not be manipulated analytically, we resort on sampling to approximate it and estimate the integrals by Monte-Carlo simulations.

Let us suppose that we can generate a sample θ 1 ,. . ., θ N distributed according to π . Obtaining such a sample itself requires careful numerical considerations discussed in the next section, but we assume for the moment that it can be obtained and show how we can estimate R ( e ; θ ) from it for a given experiment e . For that purpose, we write

for 0≤ i , j ≤ N , as a discrete estimate of the second integral in equation ( 2 ). Since

are independantly drawn from π the prior terms disappear. Moreover, the denominator is a discretization of the denominator in equation ( 2 ), and the likelihood P is supposed to be given. We have the standard estimate of ( 2 ) by an empirical average:

We see that the quantity w ij ( e ) measures how similar the observation profiles are under the two alternatives θ i and θ j . A good experiment produces dissimilar profiles and thus low values of w ij ( e ) when θ i and θ j are far appart. The resulting risk is thus reduced accordingly.

For each i and j , the quantity w ij ( e ) can in turn be estimated by Monte-Carlo simulations. For each θ i , a sample of the conditionnal distribution P ( o | θ i ; e ), denoted by o u i ( u =1,· · ·, M ) is generated. The corresponding approximation is:

which can be interpreted as a weighted likelihood of the alternative when the observation is generated according to θ i .

In most settings, generating a sample o u i involves running a deterministic model, to be performed once for each θ i , and degrading the output according to a noise model independently for each u . In our case, we used the solver proposed in [ 50 ] provided in the package [ 51 ] to simulate the ODE systems. Thus, a large number M can be used if necessary at minimal cost. Based on these samples, the approximated weights w ij M can be computed from ( 5 ), from which the expected risk of experiment e can be derived from ( 4 ).

Note that an appealing property of this scheme is that the same sample θ i can be used to evaluate all experiments. We now need to discuss how to obtain this sample.

Sampling the parameter space

Sampling the parameter space according to π k , the posterior distribution of parameters after the k -th experiment, is challenging because the likelihood function can exhibit multi-modality, plateaus and abrupt changes as illustrated in Figure 1 . Traditional sampling techniques tend to get stuck in local optima, not accounting for the diversity of high likelihood areas of the parameter space [ 52 ]. In order to speed up the convergence of sampling algorithm to high posterior density regions, we implemented a Broyden-Fletcher-Goldfarb-Shanno (BFGS) quasi-Newton optimization algorithm using finite difference approximation for gradient estimation [ 53 ] in order to identify several modes of the posterior distribution, and used these local maxima as initial values for a Metropolis Hastings sampler, combining isotropic Gaussian proposal and single parameter modifications [ 52 ]. We then use a Gaussian mixture model approximation to estimate a weighting scheme of in order to account for the initialization process when recombining samples from different modes. Annex B, given in the Additional file 1 provides computational details for this procedure.

figure 1

Log likelihood surface. Log likelihood surface for parameters living on a restricted area of a two dimensional plane. For clarity, scale is not shown. Areas with low log-likelihood correspond to dynamics that do not fit the data at all, while areas with high log-likelihood fit the data very well. The surface shows multi-modality, plateaus and abrupt jumps which makes it difficult to sample from this density. When parameters do not live on a plane, these curses have even higher effect.

The method described in Algorithm 1 is independant of the sampling scheme used. However, convergence of posterior samples is essential to ensure a good behaviour of the method. First, it is known that improper (or "flat") priors may lead to improper posterior distributions when the model contains non identifiabilities. Such issues should be avoided since MCMC based sampling schemes are known not to converge in these cases. Therefore, proper prior distributions are essential in this context and improper priors should not be used in order to avoid improper posteriors. The second important element for posterior samples is numerical convergence of the sampling scheme, usually guaranteed asymptotically. Fine tuning parameters that drive the scheme is necessary to ensure that one is close to convergence in a reasonable amount of time. To check appropriate sampling behaviour, we use a graphical heuristic. We draw ten different samples from the same posterior distribution, using different initialization seeds. For each model parameter, we compare the dispersion within each sample to the total dispersion obtained by concatenating the ten samples. This value should be close to one. Such an heuristic can be used to tune parameters of the sampler, such as sample size or proposal distribution. More details and numerical results are given in Additional file 1 : Annex B.

Enforcing regularity through the prior distribution

The prior distribution π 0 plays a crucial role at early stages of the design, as it can penalize parameters leading to dynamical behaviors that we consider unlikely. In addition to a large variance log normal prior, we considered penalizing parameters leading to non smooth time trajectories. This is done by adding to the prior log density a factor that depends on the maximum variation of time course trajectories as follows. To each parameter value θ are associated trajectories, Y i , t , which represent concentration values of the i -th species at time t . In the evaluation of the log prior density at θ , we add a term proportional to

The advantage of this is twofold. First, it is reasonable to assume that variables we do not observe in a specific design vary smoothly with time. Second, this penalization allows to avoid regions of the parameter space corresponding to very stiff systems, which are poor numerical models of reality, and which simulation are computationally demanding or simply make the solver fail. This penalty term is only used in the local optimization phase not during the Monte Carlo exploration of the posterior. The main reason for adopting such a scheme is numerical stability.

The choice of prior parameters directly affects the posterior disribution, specially when a low amount of data is available. In our experiments, the prior is chosen to be log-normal with large variance. This allows to cover a wide range of potential physical values for each parameter (from 10 -9 to 10 9 ). The weight of the regularity enforcing term has also to be determined. Since the purpose is to avoid regions corresponding to numerically unstable systems, we chose this weight to be relatively small compared to the likelihood term. In practical applications, parameters have to be chosen by considering the physical scale of quantities to be estimated. Indeed, a wrong choice of hyper parameter leads to very biased estimates at the early stages of the design.

Results and discussion

In silico network description.

In order to evaluate the relevance of our new sequential Bayesian OED strategy in the context of systems biology, we test it on an in silico network proposed in the DREAM7 Network Topology and Parameter Inference Challenge which we now describe [ 49 ]. The network, represented graphically in Figure 2 , is composed of 9 genes and its dynamics is governed by ordinary differential equations representing kinetic laws involving 45 parameters. Promoting reactions are represented by green arrows and inhibitory reactions are depicted by red arrows. For each of the 9 genes, both protein and messenger RNA are explicitly modelled and therefore the model contained 18 continuous variables. Promoter strength controls the transcription reaction and ribosomal strength controls the protein synthesis reaction. Decay of messenger RNA and protein concentrations is controlled through degradation rates. A complete description of the underlying differential equations is found in Additional file 2 : Annex A. The complete network description and implementations of integrators to simulate its dynamics are available from [ 49 ].

figure 2

Gene network for DREAM7 Challenge. Gene network for DREAM7 Network Topology and Parameter Inference Challenge. Promoting reactions are represented by green arrows and inhibitory reactions are depicted by red arrows.

Various experiments can be performed on the network producing new time course trajectories in unseen experimental conditions. An experiment consists in choosing an action to perform on the system and deciding which quantity to observe. The possible actions are

do nothing (wild type);

delete a gene (remove the corresponding species);

knock down a gene (increase the messenger RNA degradation rate by ten folds);

decrease gene ribosomal activity (decrease the parameter value by 10 folds).

These actions are coupled with 38 possible observable quantities

messenger RNA concentration for all genes, at two possible time resolutions (2 possible choices);

protein concentration for a single pair of proteins, at a single resolution (resulting in 9-8/2=36 possible choices).

Purchasing data consists in selecting an action and an observable quantities. In addition, it is possible to estimate the constants (binding affinity and hill coefficient) of one of the 13 reactions in the system. Different experiments and observable quantities have different costs, the objective being to estimate unknown parameters as accurately as possible, given a fixed initial credit budget. The cost of the possible experiments are described in Table S1 in Additional file 2 : Annex A.

For simulation purposes, we fix an unknown parameter value θ θ to control the dynamics of the systems, and the risk of an estimator is defined in terms of the loss function l ( θ , θ * ) = ∑ i = 1 p log θ i / θ i θ 2 .

The noise model used for data corruption is heteroscedastic Gaussian: given the true signal y ∈ ℝ + , the corrupted signal has the form y + z 1 + z 2 , where z 1 and z 2 are centered normal variables with standard deviation 0.1 and (0.2× y ), respectively.

Performance on a 3-gene subnetwork

In order to assess the performance of our sequential OED strategy in an easily reproducible setting, we first compare it to other strategies on a small network made of 3 genes. We take the same architecture as in Figure 2 , only considering proteins 6, 7 and 8. The resulting model has 6 variables (the mRNA and protein concentrations of the three genes) whose behavior is governed by 9 parameters. There are 50 possible experiments to choose from for this sub network: 10 perturbations (wildtype and 3 perturbations for each gene) and 5 observables (mRNA concentrations at two different time resolutions and each protein concentration at a single resolution). We compare three ways to sequentially choose experiments in order to estimate the 9 unknown parameters: (i) our new Bayesian OED strategy, including the multimodal sampling of parameter space, (ii) the criterion proposed in equation ( 13 ) in [ 27 ] together with our posterior exploration strategy, and (iii) a random experimental design, where each experiment not done yet is chosen with equal probability. The comparison of (i) and (ii) is meant to compare our strategy with a criterion that proved to be efficient in a similar setting. The comparison to (iii) is meant to assess the benefit, if any, of OED for parameter estimation in systems biology. Since all methods involve randomness, we repeat each experiment 10 times with different pseudo-random number generator seeds.

The results are presented in Figure 3 , where we show, for each of the three methods, the risk of the parameter estimation as a function of budget used. Here the risk is defined as the loss between the true parameter θ * (unknown to the method) and the estimated mean of the posterior distribution. After k rounds of experimental design, one has access to k experimental datasets which define a posterior distribution θ k from which a sample { θ i k } i = 1 N is drawn. The quantities displayed in Figure 3 are computed as

which would be the true risk that one would have to support. We first observe that the random sampling strategy has the worst risk among the three strategies, suggesting that optimizing the experiments to be made for parameter estimation outperforms a naive random choice of experiments. Second, and more importantly, the comparison between the first and second panel suggests that, given the same parameter space exploration heuristic, our proposed strategy outperforms the criterion given in [ 27 ]. It is worth noting that this criterion is part of a strategy that performed best in DREAM6 parameter estimation challenge. Although a large part of their design procedure involved human choice which we did not implement, we reproduced the part of their procedure that could be automatised. A compagnon of Figure 3 is given in Figure S3 in Additional file 1 : Annex B where we illustrate based on parameter samples how lacks of identifiability manifest themselves in a Bayesian context and how the design strategy alleviates them in terms of posterior distribution. In summary, this small experiment validates the relevance of our Bayesian OED strategy.

figure 3

Comparison of risk evolution between different strategies. Comparison of risk evolution between different strategies on a subnetwork. The figure shows the true risk at each step of the procedure, i.e. the approximate posterior distribution is compared to the true underlying parameter which is unknown during the process. The risk is computed at the center of the posterior sample. The different lines represent 10 repeats of the design procedure given the same initial credit budget and the points represent experiment purchase. The first panel represents our strategy, the second panel implements the criterion of the best performing team on DREAM6 challenge while random design consists in choosing experiments randomly.

Results on the full DREAM7 network

To illustrate the behavior of our OED strategy in a more realistic context, we then apply it to the full network of Figure 2 following the setup of the DREAM7 challenge. At the beginning of the experiment, we already have at hand low resolution mRNA time courses for the wild type system. The first experiments chosen by the method are wild-type protein concentration time courses for all genes. The detailed list of purchased experiments is found Table S2 in Additional file 2 : Annex A. This makes sense since we have enormous uncertainty about proteins time courses, given that we do not know anything about them. Once these protein time series are purchased, the suggestion for the next experiment to carry out is illustrated in Table 1 . Interestingly, the perturbations with the lowest risk are related to gene 7 which is on the top of the cascade (see Figure 2 ). Moreover it seemed obvious from Table 1 that we have to observe protein 8 concentration. Indeed, Figure 4 shows that there is a lot of uncertainty about protein 8 evolution when we remove gene 7.

figure 4

Trajectories from posterior sample. Corresponds to Table 1 figures. We plot trajectories from our posterior sample (protein 8 concentration was divided by 2 and we do not plot concentrations higher than 100). The quantities with the highest variability are protein 8 and 3 concentrations. This is consistent with the estimated risks in Table 1 . There is quite a bit of uncertainty in protein 5 concentration, however this is related to protein 8 uncertainty as protein 8 is an inhibitor of protein 5. Moreover, mRNA concentration have much lower values and are not as informative as proteins concentrations. Red dots shows the data we purchased for this experiment after seeing these curve and in accordance with results in Table 1 .

Moreover, our criterion determines that it is better to observe protein 3 than protein 5, which makes sense since the only protein which affects protein 5 evolution is protein 8 (see Figure 2 ). Therefore uncertainty about protein 5 time course is tightly linked to protein 8 time course, and observing protein 3 brings more information than observing protein 5. This might not be obvious when looking at the graph in Figure 4 and could not have been foreseen by a method that considers uncertainty about each protein independently. At this point, we purchase protein 3 and 8 time courses for gene 7 deletion experiment and highlight in red in Figure 4 the profiles of proteins 3 and 8 obtained from the system.

In addition to parameter estimation, one may be interested in the ability of the model with the inferred parameters to correctly simulate time series under different experimental conditions. Figure 5 represents a sample from the posterior distribution after all credits have been spent (unseen experiment description is given in Table S3 Additional file 2 : Annex A). Both parameter values and protein time course for the unseen experiment are presented.

figure 5

Comparison of parameter and trajectory variability. Comparison of parameter variability and time course trajectory variability. This is a sample from the posterior distribution after spending all the credits in the challenge. The top of the figure shows parameter values on log scale, while the bottom shows prediction of protein time courses for an unseen experiment. The range of some parameter values is very wide while all these very different values lead to very similar protein time course predictions.

Some parameters, like p _ d e g r a d a t i o n _ r a t e or p r o 3_ s r e n g h t , clearly concentrate around a single value while others, like p r o 1_ s t r e n g t h or p r o 2_ s t r e n g t h , have very wide ranges with multiple accumulation points. Despite this variability in parameter values, the protein time course trajectories are very similar. It appears that protein 5 time course is less concentrated than the two others. This is due to the hetroscedasticity of the noise model which was reflected in the likelihood. Indeed, the noise model is Gaussian with standard deviation increasing with the value of the corresponding concentration. Higher concentrations are harder to estimate due to larger noise standard deviation.

Computational systems biology increasingly relies on the heavy use of computational resources to improve the understanding of the complexity underlying cell biology. A widespread approach in computational systems biology is to specify a dynamical model of the biological process under investigation based on biochemical knowledge, and consider that the real system follows the same dynamics for some kinetic parameter values. Recent reports suggest that this has benefits in practical applications ( e.g. [ 54 ]). Systematic implementations of the approach requires to deal with the fact that most kinetic parameters are often unknown, raising the issue of estimating these parameters from experimental data as efficiently as possible. An obvious sanity check is to recover kinetic parameters from synthetic data where dynamic and noise model are well specified, which is already quite a challenge.

In this paper we proposed a new general Bayesian OED strategy, and illustrated its relevance on an in silico biological network. The method takes advantage of the Bayesian framework to sequentially choose experiments to be performed, in order to estimate these parameters subject to cost constraints. The method relies on a single numerical criterion and does not depend on a specific instance of this problem. This is in our opinion a key point in order to reproducibly be able to deal with large scale networks of size comparable to of a cell for example. Experimental results suggest that the strategy has the potential to support experimental design in systems biology.

As noted by others [ 11 ],[ 12 ],[ 15 ]-[ 17 ], the approach focusing on kinetic parameter estimation is questionable. We also give empirical evidence that very different parameter values can produce very similar dynamical behaviors, potentially leading to non-identifiability issues. Moreover, focusing on parameter estimation supposes that the dynamical model represents the true underlying chemical process. In some cases, this might simply be false. For example, hypotheses underlying the law of mass action are not satisfied in the gene transcription process. However, simplified models might still be good proxies to characterize dynamical behaviors we are interested in. The real problem of interest is often to reproduce the dynamics of a system in terms of observable quantities, and to predict the system behavior for unseen manipulations. Parameters can be treated as latent variables which impact the dynamics of the system but cannot be observed. In this framework, the Bayesian formalism described here is well suited to tackle the problem of experimental design.

The natural continuity of this work is to adapt the method to treat larger problems. This raises computational issues and requires to develop numerical methods that scale well with the size of the problem. Sampling strategies that adapt to the local geometry and to multimodal aspects of the posterior, such as described e.g. in [ 55 ],[ 56 ] are interesting directions to investigae in this context. The main bottlenecks are the cost of simulating large dynamical systems, and the need for large sample size in higher dimension for accurate posterior estimation. Posterior estimation in high dimensions is known to be hard and is an active subject of research. Although our Bayesian OED criterion is independent of the model investigated, it is likely that a good sampling strategy to implement may benefit from specific tuning in order to perform well on specific problem instances. As for reducing the computational burden of simulating large dynamical systems, promising research directions are parameter estimation methods that do not involve dynamical system simulation such as [ 57 ] or differential equation simulation methods that take into account both parameter uncertainty and numerical uncertainty such as the probabilistic integrator of [ 58 ].

Availability of supporting data

An R R package that allows to reproduce our results and simulations is available at the following URL: https://doi.org/cran.r-project.org/package=pauwels2014 .

Additional files

Kitano H: Computational systems biology. Nature 2002, 420:206-210.

Article   CAS   PubMed Central   Google Scholar  

Tyson JJ, Chen KC, Novak B: Sniffers, buzzers, toggles and blinkers: dynamics of regulatory and signaling pathways in the cell. Curr Opin Cell Biol. 2003, 15 (2): 221–231.

Wilkinson DJ: Stochastic modelling for quantitative description of heterogeneous biological systems. Nat Rev Genet. 2009, 10 (2): 122–133.

Barillot E, Calzone L, Vert J-P, Zynoviev A: Computational Systems Biology of Cancer. 2012, CRC Press, London

Book   Google Scholar  

Chen KC, Csikasz-Nagy A, Gyorffy B, Val J, Novak B, Tyson JJ: Kinetic analysis of a molecular model of the budding yeast cell cycle. Mol Biol Cell. 2000, 11 (1): 369–391.

Hoffmann A, Levchenko A, Scott ML, Baltimore D: The i k b-nf- k b signaling module: temporal control and selective gene activation. Science. 2002, 298 (5596): 1241–1245.

Lipniacki T, Paszek P, Brasier AR, Luxon B, Kimmel M: Mathematical model of nf-kappab regulatory module. J Theor Biol. 2004, 228 (2): 195–215.

Schoeberl B, Eichler-Jonsson C, Gilles ED, Müller G: Computational modeling of the dynamics of the map kinase cascade activated by surface and internalized egf receptors. Nat Biotechnol. 2002, 20 (4): 370–375.

Article   PubMed Central   Google Scholar  

Fussenegger M, Bailey J, Varner J: A mathematical model of caspase function in apoptosis. Nat Biotechnol 2000, 18:768–774.

Jaqaman K, Danuser G: Linking data to models: data regression. Nat Rev Mol Cell Biol. 2006, 7 (11): 813–819.

Battogtokh D, Asch DK, Case ME, Arnold J, Schuttler H-B: An ensemble method for identifying regulatory circuits with special reference to the qa gene cluster of neurospora crassa. Proc Natl Acad Sci U S A. 2002, 99 (26): 16904–16909.

Brown KS, Sethna JP: Statistical mechanical approaches to models with many poorly known parameters. Phys Rev E 2003, 68:021904.

Brown KS, Hill CC, Calero GA, Myers CR, Lee KH, Sethna JP, Cerione RA: The statistical mechanics of complex signaling networks: nerve growth factor signaling. Phys Biol. 2004, 1 (3–4): 184–195.

Article   CAS   Google Scholar  

Waterfall JJ, Casey FP, Gutenkunst RN, Brown KS, Myers CR, Brouwer PW, Elser V, Sethna JP: Sloppy-model universality class and the Vandermonde matrix. Phys Rev Lett. 2006, 97 (15): 150601

Article   Google Scholar  

Achard P, De Schutter E: Complex parameter landscape for a complex neuron model. PLoS Comput Biol. 2006, 2 (7): 94

Gutenkunst RN, Waterfall JJ, Casey FP, Brown KS, Myers CR, Sethna JP: Universally sloppy parameter sensitivities in systems biology models. PLoS Comput Biol. 2007, 3 (10): 1871–1878.

Piazza M, Feng X-J, Rabinowitz JD, Rabitz H: Diverse metabolic model parameters generate similar methionine cycle dynamics. J Theor Biol. 2008, 251 (4): 628–639.

Bellman R, Åström KJ: On structural identifiability. Math Biosci. 1970, 7 (3): 329–339.

Cobelli C, DiStefano JJ: Parameter and structural identifiability concepts and ambiguities: a critical review and analysis. Am J Physiol Regul Integr Comp Physiol. 1980, 239 (1): R7–R24.

Villaverde AF, Banga JR: Reverse engineering and identification in systems biology: strategies, perspectives and challenges. J R Soc Interface 2014, 11(91).

Mendes P, Kell D: Non-linear optimization of biochemical pathways: applications to metabolic engineering and parameter estimation. Bioinformatics. 1998, 14 (10): 869–883.

Moles CG, Mendes P, Banga JR: Parameter estimation in biochemical pathways: a comparison of global optimization methods. Genome Res. 2003, 13 (11): 2467–2474.

Fernández Slezak D, Suárez C, Cecchi GA, Marshall G, Stolovitzky G: When the optimal is not the best: parameter estimation in complex biological models. PLoS One. 2010, 5 (10): 13283

Kreutz C, Timmer J: Systems biology: experimental design. FEBS J. 2009, 276 (4): 923–942.

Cho K-H, Shin S-Y, Kolch W, Wolkenhauer O: Experimental design in systems biology, based on parameter sensitivity analysis using a monte carlo method: a case study for the TNF α -mediated NF- k B signal transduction pathway. Simulation. 2003, 79 (12): 726–739.

Feng X-J, Rabitz H: Optimal identification of biochemical reaction networks. Biophys J. 2004, 86 (3): 1270–1281.

Steiert B, Raue A, Timmer J, Kreutz C: Experimental design for parameter estimation of gene regulatory networks. PLoS One. 2012, 7 (7): 40052

Chaloner K, Verdinelli I: Bayesian experimental design: a review. Stat Sci. 1995, 10 (3): 273–304.

Lindley DV: On a measure of the information provided by an experiment. Ann Math Stat. 1956, 27 (4): 986–1005.

Bandara S, Schlöder JP, Eils R, Bock HG, Meyer T: Optimal experimental design for parameter estimation of a cell signaling model. PLoS Comput Biol. 2009, 5 (11): 1000558

Transtrum MK, Qiu P: Optimal experiment selection for parameter estimation in biological differential equation models. BMC Bioinformatics 2012, 13:181.

Faller D, KlingMüller U, Timmer J: Simulation methods for optimal experimental design in systems biology. Simulation 2003, 79:717–725.

Kutalik Z, Cho K-H, Wolkenhauer O: Optimal sampling time selection for parameter estimation in dynamic pathway modeling. Biosystems. 2004, 75 (1–3): 43–55.

Gadkar KG, Gunawan R, Doyle FJ 3rd: Iterative approach to model identification of biological networks. BMC Bioinformatics 2005, 6:155.

Balsa-Canto E, Alonso AA, Banga JR: Computational procedures for optimal experimental design in biological systems. IET Syst Biol. 2008, 2 (4): 163–172.

Hagen DR, White JK, Tidor B: Convergence in parameters and predictions using computational experimental design. Interface Focus. 2013, 3 (4).

Kramer A, Radde N: Towards experimental design using a bayesian framework for parameter identification in dynamic intracellular network models. Procedia Comput Sci. 2010, 1 (1): 1645–1653.

Liepe J, Filippi S, Komorowski M, Stumpf MPH: Maximizing the information content of experiments in systems biology. PLoS Comput Biol. 2013, 9 (1): 1002888

Casey FP, Baird D, Feng Q, Gutenkunst RN, Waterfall JJ, Myers CR, Brown KS, Cerione RA, Sethna JP: Optimal experimental design in an epidermal growth factor receptor signalling and down-regulation model. IET Syst Biol. 2007, 1 (3): 190–202.

Weber P, Kramer A, Dingler C, Radde N: Trajectory-oriented bayesian experiment design versus Fisher A-optimal design: an in depth comparison study. Bioinformatics. 2012, 28 (18): 535–541.

Vanlier J, Tiemann CA, Hilbers PAJ: A bayesian approach to targeted experiment design. Bioinformatics. 2012, 28 (8): 1136–1142.

Kremling A, Fischer S, Gadkar K, Doyle FJ, Sauter T, Bullinger E, Allgöwer F, Gilles ED: A benchmark for methods in reverse engineering and model discrimination: problem formulation and solutions. Genome Res. 2004, 14 (9): 1773–1785.

Apgar JF, Toettcher JE, Endy D, White FM, Tidor B: Stimulus design for model selection and validation in cell signaling. PLoS Comput Biol. 2008, 4 (2): e30

Busetto AG, Hauser A, Krummenacher G, Sunnåker M, Dimopoulos S, Ong CS, Stelling J, Buhmann JM: Near-optimal experimental design for model selection in systems biology. Bioinformatics. 2013, 29 (20): 2625–2632.

Skanda D, Lebiedz D: An optimal experimental design approach to model discrimination in dynamic biochemical systems. Bioinformatics. 2010, 26 (7): 939–945.

Roy N, McCallum A: Toward optimal active learning through sampling estimation of error reduction. Proceedings of the Eighteenth International Conference on Machine Learning. 2001, Morgan Kaufmann Publishers Inc, San Francisco, 441–448.

Google Scholar  

Dialogue for Reverse Engineering Assessments and Methods (DREAM) website. []. Accessed 2013–1228., [ https://doi.org/www.the-dream-project.org ]

DREAM6 Estimation of Model Parameters Challenge website. [] Accessed 2013–1228., [ https://doi.org/www.the-dream-project.org/challenges/dream6-estimation-model-parameters-challenge ]

DREAM7 Estimation of Model Parameters Challenge website. []. Accessed 2013–1228., [ https://doi.org/www.the-dream-project.org/challenges/network-topology-and-parameter-inference-challenge ]

Bogacki P, Shampine LF: A 3(2) pair of Runge — Kutta formulas. Appl Math Lett 1989, 2:321–325.

Soetaert K, Petzoldt T, Setzer RW: Solving differential equations in R: package deSolve. J Stat Softw. 2010, 33 (9): 1–25.

Andrieu C, De Freitas N, Doucet A, Jordan MI: An introduction to MCMC for machine learning. Mach Learn. 2003, 50 (1–2): 5–43.

Nocedal J, Wright S: Numerical Optimization. 2006, Springer, New York

Karr JR, Sanghvi JC, Macklin DN, Gutschow MV, Jacobs JM, Bolival B, Assad-Garcia N, Glass JI, Covert MW: A whole-cell computational model predicts phenotype from genotype. Cell. 2012, 150 (2): 389–401.

Girolami M, Calderhead B: Riemann manifold langevin and hamiltonian monte carlo methods. J Roy Stat Soc B Stat Meth. 2011, 73 (2): 123–214.

Calderhead B, Girolami M: Estimating Bayes factors via thermodynamic integration and population MCMC. Comput Stat Data Anal. 2009, 53 (12): 4028–4045.

Calderhead B, Girolami M, Lawrence ND: Accelerating Bayesian inference over nonlinear differential equations with Gaussian processes. In Adv. Neural. Inform. Process Syst : MIT Press; 2008:217–224.

Chkrebtii O, Campbell DA, Girolami MA, Calderhead B: Bayesian uncertainty quantification for differential equations 2013. Technical Report 1306.2365, arXiv.

Download references

Acknowledgements

The authors would like to thank Gautier Stoll for insightful discussions. This work was supported by the European Research Council (SMAC-ERC-280032). Most of this work was carried out during EP’s PhD at Mines ParisTech.

Author information

Authors and affiliations.

CNRS, LAAS, 7 Avenue du Colonel Roche, Toulouse, F-31400, France

Edouard Pauwels

Univ de Toulouse LAAS, Toulouse, F-31400, France

MINES ParisTech, PSL-Research University, CBIO-Centre for Computational Biology, 35 rue Saint-Honoré, Fontainebleau, 77300, France

Christian Lajaunie & Jean-Philippe Vert

Institut Curie, 26 rue d’Ulm, F-75248, Paris, France

INSERM U900, Paris, F-75248, France

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to Edouard Pauwels .

Additional information

Competing interests.

The authors declare that they have no competing interests.

Authors' contributions

EP was responsible for the implementation and drafting the manuscript. CL and JPV supervised the work. All three authors took part in the design of the method and the writing of the manuscript. All authors read and approved the final manuscript.

Electronic supplementary material

12918_2014_102_moesm1_esm.pdf.

Additional file 1: Annex B. Supplementary details regarding the sampling strategy used in our numerical experiments. The note also contains diagnosis information and marginal distribution samples to illustrate the efficacy of the sampling strategy in the setting of this paper. (PDF 229 KB)

12918_2014_102_MOESM2_ESM.pdf

Additional file 2: Annex A. PDF file. Description of the DREAM7 challenge network represented in Figure 2 and experimental design setting. (PDF 102 KB)

Authors’ original submitted files for images

Below are the links to the authors’ original submitted files for images.

Authors’ original file for figure 1

Authors’ original file for figure 2, authors’ original file for figure 3, authors’ original file for figure 4, authors’ original file for figure 5, authors’ original file for figure 6, rights and permissions.

This article is distributed under the terms of the Creative Commons Attribution 4.0 International License ( http://creativecommons.org/licenses/by/4.0/ ), which permits unrestricted use, distribution, and reproduction in any medium, provided you give appropriate credit to the original author(s) and the source, provide a link to the Creative Commons license, and indicate if changes were made. The Creative Commons Public Domain Dedication waiver ( http://creativecommons.org/publicdomain/zero/1.0/ ) applies to the data made available in this article, unless otherwise stated.

Reprints and permissions

About this article

Cite this article.

Pauwels, E., Lajaunie, C. & Vert, JP. A Bayesian active learning strategy for sequential experimental design in systems biology. BMC Syst Biol 8 , 102 (2014). https://doi.org/10.1186/s12918-014-0102-6

Download citation

Received : 10 February 2014

Accepted : 14 August 2014

Published : 01 December 2014

DOI : https://doi.org/10.1186/s12918-014-0102-6

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • Systems biology
  • Kinetic parameter estimation
  • Active learning
  • Bayesian experimental design

BMC Systems Biology

ISSN: 1752-0509

experimental methods in systems biology

Loading metrics

Open Access

Peer-reviewed

Research Article

Model Selection in Systems Biology Depends on Experimental Design

Affiliation Centre for Integrative Systems Biology at Imperial College London, London, United Kingdom

* E-mail: [email protected]

  • Daniel Silk, 
  • Paul D. W. Kirk, 
  • Chris P. Barnes, 
  • Tina Toni, 
  • Michael P. H. Stumpf

PLOS

  • Published: June 12, 2014
  • https://doi.org/10.1371/journal.pcbi.1003650
  • Reader Comments

Figure 1

Experimental design attempts to maximise the information available for modelling tasks. An optimal experiment allows the inferred models or parameters to be chosen with the highest expected degree of confidence. If the true system is faithfully reproduced by one of the models, the merit of this approach is clear - we simply wish to identify it and the true parameters with the most certainty. However, in the more realistic situation where all models are incorrect or incomplete, the interpretation of model selection outcomes and the role of experimental design needs to be examined more carefully. Using a novel experimental design and model selection framework for stochastic state-space models, we perform high-throughput in-silico analyses on families of gene regulatory cascade models, to show that the selected model can depend on the experiment performed. We observe that experimental design thus makes confidence a criterion for model choice, but that this does not necessarily correlate with a model's predictive power or correctness. Finally, in the special case of linear ordinary differential equation (ODE) models, we explore how wrong a model has to be before it influences the conclusions of a model selection analysis.

Author Summary

Different models of the same process represent distinct hypotheses about reality. These can be decided between within the framework of model selection, where the evidence for each is given by their ability to reproduce a set of experimental data. Even if one of the models is correct, the chances of identifying it can be hindered by the quality of the data, both in terms of its signal to measurement error ratio and the intrinsic discriminatory potential of the experiment undertaken. This potential can be predicted in various ways, and maximising it is one aim of experimental design. In this work we present a computationally efficient method of experimental design for model selection. We exploit the efficiency to consider the implications of the realistic case where all models are more or less incorrect, showing that experiments can be chosen that, considered individually, lead to unequivocal support for opposed hypotheses.

Citation: Silk D, Kirk PDW, Barnes CP, Toni T, Stumpf MPH (2014) Model Selection in Systems Biology Depends on Experimental Design. PLoS Comput Biol 10(6): e1003650. https://doi.org/10.1371/journal.pcbi.1003650

Editor: Burkhard Rost, Tum, Germany

Received: May 30, 2013; Accepted: April 10, 2014; Published: June 12, 2014

Copyright: © 2014 Silk et al. This is an open-access article distributed under the terms of the Creative Commons Attribution License , which permits unrestricted use, distribution, and reproduction in any medium, provided the original author and source are credited.

Funding: This work was funded by the Biotechnology and Biological Science Research Council ( www.bbsrc.ac.uk ) grant BB/K003909/1 (to DS and MPHS, the Human Frontiers Science Programme (ww.hfsp.org) grant RG0061/2011 (to PDWK and MPHS), and a Wellcome Trust-MIT fellowship ( www.wellcome.ac.uk ) to TT. The funders had no role in study design, data collection and analysis, decision to publish, or preparation of the manuscript.

Competing interests: The authors have declared that no competing interests exist.

Introduction

Mathematical models provide a rich framework for biological investigation. Depending upon the questions posed, the relevant existing knowledge and alternative hypotheses may be combined and conveniently encoded, ready for analysis via a wealth of computational techniques. The consequences of each hypothesis can be understood through the model behaviour, and predictions made for experimental validation. Values may be inferred for unknown physical parameters and the actions of unobserved components can be predicted via model simulations. Furthermore, a well-designed modelling study allows conclusions to be probed for their sensitivity to uncertainties in any assumptions made, which themselves are necessarily made explicit.

While the added value of a working model is clear, how to create one is decidedly not. Choosing an appropriate formulation (e.g. mechanistic, phenomenological or empirical), identifying the important components to include (and those that may be safely ignored), and defining the laws of interaction between them remains highly challenging, and requires a combination of experimentation, domain knowledge and, at times, a measure of luck. Even the most sophisticated models will still be subject to an unknown level of inaccuracy – how this affects the modelling process, and in particular experimental design for Bayesian inference, will be the focus of this study.

Both the time and financial cost of generating data, and a growing understanding of the data dependancy of model and parameter identifiability [1] , [2] , has driven research into experimental design. In essence, experimental design seeks experiments that maximise the expected information content of the data with respect to some modelling task. Recent developments include the work of Liepe et. al [2] that builds upon existing methods [3] – [8] , by utilising a sequential approximate Bayesian computation framework to choose the experiment that maximises the expected mutual information between prior and posterior parameter distributions. In so doing, they are able to optimally narrow the resulting posterior parameter or predictive distributions, incorporate preliminary experimental data and provide sensitivity and robustness analyses. In a markedly different approach, Apgar et. al [8] use control theoretic principles to distinguish between competing models; here the favoured model is that which is best able to inform a controller to drive the experimental system through a target trajectory.

experimental methods in systems biology

  • PPT PowerPoint slide
  • PNG larger image
  • TIFF original image

experimental methods in systems biology

https://doi.org/10.1371/journal.pcbi.1003650.g001

The contributions of this article are threefold; firstly, we extend a promising and computationally efficient experimental design framework for model selection to the stochastic setting, with non-Gaussian prior distributions; secondly, we utilise this efficiency to explore the robustness of model selection outcomes to experimental choices; and finally, we observe that experimental design can give rise to levels of confidence in selected models that may be misleading as a guide to their predictive power or correctness. The latter two points are undertaken via high-throughput in-silico analyses (at a scale completely beyond the Monte Carlo based approaches mentioned above) on families of gene regulatory cascade models and various existing models of the JAK STAT pathway.

Identifying crosstalk connections between signalling pathways

We first illustrate the experimental design and model selection framework in the context of crosstalk identification. After observing how the choice of experiment can be crucial for a positive model selection outcomes, the example will be used to illustrate and explore the inconsistency of selection between misspecified models.

experimental methods in systems biology

https://doi.org/10.1371/journal.pcbi.1003650.g002

experimental methods in systems biology

https://doi.org/10.1371/journal.pcbi.1003650.g003

experimental methods in systems biology

https://doi.org/10.1371/journal.pcbi.1003650.g004

experimental methods in systems biology

The robustness of model selection to choice of experiment

experimental methods in systems biology

The results for data sets of size 1 and 8 are illustrated in Figure 4c and 4d as heatmaps of posterior probabilities of the first model, and show that the vast majority of the space of experiments is split into distinct regions of high, low and equal probability for each model. In the case of a single time point, most of the explored experiment subspace is found to be uninformative, with the data providing equal support for each model. Three other distinct regions are identified, of which two show decisive support (on the Jeffreys scale) for the first model, and one for which the second model is chosen decisively. In other words, by varying the experimental conditions an unequivocal choice (in isolation) for either model can be obtained. As more data points are considered, the uninformative region grows smaller, but regions of decisive support for each model remain. Interestingly, these regions are located in distinctly different places for single or multiple time points, although they remain similar for 2 or more time points. This reflects the added value of time series experiments – the marginal likelihoods now balance the ability of the models to reproduce each time point, with their ability to capture the autocorrelation of the time series.

experimental methods in systems biology

The top right plot shows posterior model probabilities obtained using MultiNest. The necessarily course grained results match those obtained by the UT in figure 4c . Each of the other plots compare UT approximations to the prior predictive distributions with Monte Carlo approximations using samples of size 10000, for different experimental conditions indicated by arrows. The red and blue lines correspond to UT approximations (using a single Gaussian component) for model 1 and 2 respectively. The dotted line indicates the data simulated from the true model.

https://doi.org/10.1371/journal.pcbi.1003650.g005

JAK-STAT signalling

experimental methods in systems biology

https://doi.org/10.1371/journal.pcbi.1003650.g006

experimental methods in systems biology

https://doi.org/10.1371/journal.pcbi.1003650.g007

Both these case studies make it clear that under the realistic assumption that all models are more or less incorrect, model selection outcomes can be sensitive to the choice of experiment. This observation has particular importance for studies that treat models as competing hypotheses that are decided between using experimental data; it is quite possible that if different experiments are undertaken, the conclusions drawn will also be different. In particular, the confidence calculated for such a conclusion (using the Jeffreys scale or another measure) can be misleading as a guide to how correct or predictive a model is ( Figure 8a ); in both the examples studied here, conditions exist such that any of the competing models can score a 'decisive' selection. The model selection outcome and associated confidence must therefore be strictly interpreted, as only increasing the odds of one model (with respect to others) for the data gathered under the specific experimental conditions.

thumbnail

https://doi.org/10.1371/journal.pcbi.1003650.g008

In light of this observation, the role of experimental design may need to be examined further. Since different models can be selected depending on the experiment undertaken, the use of experimental design will necessarily lead to choosing the model which, for some 'optimal' experiment, has the highest possible predicted level of confidence i.e. experimental design implicitly makes confidence a selection criterion. Is it misleading to claim high confidence in a model selection result when the models have been set up (by extensions to mimic the optimal experiment) for this purpose? Is a bias introduced into the inference via experiment design? In the context of experiment design for parameter estimation, MacKay suggests this is not a problem [17] , stating that Bayesian inference depends only on the data collected, and not on other data that could have been gathered but was not. Our situation here is different since we consider changes not only to the data collection procedure, but also the data generation process and in turn the competing models themselves. It seems plausible that some models will gain or lose more flexibility than others with regards to fitting data for a particular choice of experiment. Even if the actual model selection is not biased, the confidence we associate with it will scale with the optimality of the experiment. After performing the optimal experiment, should there be any surprise that the selected model seems to have high support from the data? We feel these questions need further investigation.

Measuring sensitivity to model inaccuracies

In practical terms, the important question seems to be: how wrong does the model structure (or parameter values) have to be before the less predictive model (or that which captures less about the true system) is chosen? Clearly the answer is sensitive to the system and models under study, and moreover, the issue of how to compare the size of different structural inaccuracies is non trivial. Here, as a first attempt, we limit ourselves to considering the simple case of parameter inaccuracies in linear ODE models.

experimental methods in systems biology

https://doi.org/10.1371/journal.pcbi.1003650.g009

experimental methods in systems biology

In real applications, where the true model is unknown and more complex, it may not be possible to tell whether a conclusion is an artefact of model inaccuracies, even when the truth of the conclusion itself can be tested by direct experimental measurement. However, the type of analysis undertaken here at least gives a measure of robustness for the conclusion to a range of model inaccuracies. Unfortunately, this remains difficult to implement in a more general setting – for example, in climatology, where the accepted method of coping with structural uncertainty is through the use of large ensembles of similar models produced by various research groups [18] , a luxury that cannot be afforded on the scale of the most ambitious systems biology projects. While the practical challenges of dealing with large numbers of models is somewhat overcome by the model selection algorithm described above, a harder conceptual problem exists of how to define perturbations to more complicated classes of model, and to compare their strengths.

Finally, the example also highlights the difficulty of testing a hypothesis that represents only part of a model. The study shows that the implicit assumption that the base model is accurate, is not necessarily benign, and can affect any conclusions drawn – a result that is borne out by the logical principle that from a false statement, anything is provable.

The scale of the analyses detailed above, comprising thousands of marginal likelihood computations, requires extreme computational efficiency. Indeed it is completely beyond Monte Carlo based methods such as that recently developed by Liepe et al. [2] , which are limited to exploring small sets of models and experiments. Here, the efficiency was obtained by using the unscented transform for propagating Gaussian mixture distributions through non-linear functions. Further computational savings can be made by exploiting the highly parallelizable nature of Flassig and Sundmacher's method [9] , which we have extended for use with mixture distributed priors and stochastic state space models.

This efficiency has allowed us to explore model selection problems involving relatively large numbers of models and experiments, and investigate the robustness of model selection results to both changes in experimental conditions and inaccuracies in the models. Results from the latter two studies illustrate some common, but often ignored, pitfalls associated with modelling and inference. Firstly, we show that the conclusions of a model selection analysis can change depending on the experiment undertaken. Related to this, we observe that confidence in such a conclusion is not a good estimator of the predictive power of a model, or the correctness of the model structure. Further we note that the use of experimental design in this context maximises the expected discriminatory information available, and implicitly makes confidence in the outcome a criterion for model selection. In the future we intend to investigate the desirability of this property and how it affects the interpretation of the confidence associated with model selection outcomes.

At the heart of these issues is a lack of understanding of the implications of model (or parameter) inaccuracies. Often improved fits to data or better model predictions are interpreted as evidence that more about the true system is being captured. This assumption underlines a guiding paradigm of systems biology [19] , where a modelling project is ideally meant to be a cycle of model prediction, experimental testing and subsequent data inspired model/parameter improvement. However, it is possible that improved data fitting and predictive power (although desirable in their own right) can be achieved by including more inaccuracies in the model. In the context of parameter estimation, this concept of local optima is widely known, and their avoidance is a challenge when performing any non-trivial inference. One simple method to do so is to include random perturbations in the inference, in order to 'kick' the search out of a local optimum. Perhaps a similar strategy might be included in the modelling paradigm; by performing random experiments, or adding or removing interactions in a model structure, data might be gathered or hypotheses generated that allows a leap to be made to a more optimal solution.

While we have been concerned solely with the statistical setting, it is reasonable to expect similar results can be found for alternative model discrimination approaches e.g the use of Semidefinite programming to establish lower bounds on the discrepancy between candidate models and data [20] . Here the particular subset of models that are invalidated will be dependent upon the experiment undertaken. However, emphasis on invalidating wrong models instead of evaluating the relative support for each at least reduces the temptation for extrapolated and, perhaps, false conclusions.

George E. P. Box famously stated that 'Essentially, all models are wrong, but some are useful'. Here we would add that if nothing else, models provide a natural setting for mathematicians, engineers and physicists to explore biological problems, exercise their own intuitions, apply theoretical techniques, and ultimately generate novel hypotheses. Whether the hypotheses are correct or not, the necessary experimental checking will reveal more about the biology.

Materials and Methods

The unscented transform.

experimental methods in systems biology

Unscented model selection

experimental methods in systems biology

Experimental design

experimental methods in systems biology

We assume that there is overlap between the system observables appearing in each model so that experiments that allow model comparison can be designed.

experimental methods in systems biology

Experimental design as an optimisation problem

experimental methods in systems biology

Acknowledgments

We would like to thank the members of the Theoretical Systems Biology Group at Imperial College London for fruitful discussions. We dedicate this paper to Joe Silk.

Author Contributions

Conceived and designed the experiments: DS MPHS. Performed the experiments: DS. Analyzed the data: DS CPB. Contributed reagents/materials/analysis tools: DS CPB PDWK TT MPHS. Wrote the paper: DS CPB MPHS.

  • View Article
  • Google Scholar
  • 18. Team CW (2010) Good practice guidance paper on assessing and combining multi model climate projections. In: IPCC Expert Meeting on Assessing and Combining Multi Model Climate Projections. p. 1.
  • 21. Julier SJ The scaled unscented transformation. In: Proceedings of the 2002 American Control Conference. American Automatic Control Council. pp. 4555–4559.
  • 23. Jeffreys H (1961) Theory of Probability. 3rd ed. Oxford: The Clarendon Press.
  • 27. Faubel F, McDonough J (2009) The split and merge unscented Gaussian mixture filter. Signal Processing Letters.
  • 30. Kristan M, Leonardis A (2010) Multivariate online kernel density estimation. In: Computer Vision Winter Workshop. pp. 77–86.

The systems biology of signaling pathways

  • Cell Biophysics
  • Published: 13 May 2016
  • Volume 61 , pages 78–84, ( 2016 )

Cite this article

experimental methods in systems biology

  • I. N. Lavrik 1 , 2 &
  • M. G. Samsonova 3  

126 Accesses

2 Citations

Explore all metrics

Signaling networks play the central role in the regulation of processes in a single cell and in the entire body. A recent breakthrough in technologies for systems biology, which combine experimental and mathematical methods, permits scientists to model signaling pathways in an individual cell and in cell populations. This approach provides new information on mechanisms that regulate a variety of biological processes. Here we discuss the mathematical formalisms that are applied to signaling pathway modeling and relevant experimental methods.

This is a preview of subscription content, log in via an institution to check access.

Access this article

Subscribe and save.

  • Get 10 units per month
  • Download Article/Chapter or eBook
  • 1 Unit = 1 Article or 1 Chapter
  • Cancel anytime

Price includes VAT (Russian Federation)

Instant access to the full article PDF.

Rent this article via DeepDyve

Institutional subscriptions

Similar content being viewed by others

experimental methods in systems biology

Modeling Signaling Networks with Different Formalisms: A Preview

experimental methods in systems biology

Introduction to Systems Biology

experimental methods in systems biology

Derivation and Use of Mathematical Models in Systems Biology

Abbreviations.

mitogen-activated protein kinase

fluorescence resonance energy transfer

N. Bluthgen and H. Herzel, in Proc. 2nd Workshop on Computation of Biochemical Pathways and Genetic Networks (Logos-Verlag, Berlin, 2001), pp. 55–62.

Google Scholar  

S. L. Spencer and P. K. Sorger, Cell, 144 , 926 (2011).

Article   Google Scholar  

A. H. Singh, D. M. Wolf, P. Wang, and A. P. Arkon, Proc. Natl. Acad. Sci. U. S. A. 105 , 7500 (2008).

Article   ADS   Google Scholar  

D. Hyduke and B. O. Palsson, Nature Rev. Genet. 11 , 297 (2010).

B. N. Kholodenko, Nature Rev. Mol. Cell Biol. 7 , 165 (2006).

J. Huard, S. Mueller, E. D. Gilles, et al., FEBS J. 279 , 3290 (2012).

J. Saez-Rodriguez, L. Simeoni, J. A. Lindquist, et al., PLoS Comput. Biol. 3 (8), e163 (2007).

Article   ADS   MathSciNet   Google Scholar  

S. A. Kauffman, J. Theor. Biol. 22 , 437 (1969).

Article   MathSciNet   Google Scholar  

S. A. Kauffman, The Origins of Order: Self-Organization and Selection in Evolution (Oxford Univ. Press, Oxford, 1993).

R. Thomas, J. Theor. Biol. 42 (3), 563 (1973).

R. Thomas and R. D’Ari, Biological Feedback (CRC Press, 1990).

MATH   Google Scholar  

R. Thomas, D. Thieffry, and M. Kaufman, Bull. Math. Biol, 57 , 247 (1995).

A. Mbodj, G. Junion, C. Brun, et al., Mol. Biosyst. 9 (9), 2248 (2013).

A. Faure, A. Naldi, F. Lopez, et al., Mol. Biosyst. 5 (12), 1787 (2009).

M. Chaves, R. Albert, and E. D. Sontag, J. Theor. Biol. 235 , 431 (2005).

A. Naldi, Lect. Notes Comput. Sci. 4695 , 233 (2007).

A. Garg, A. Di Cara, I. Xenarios, et al., Bioinformatics 24 , 1917 (2008).

E. Klipp, W. Liebermeister, C. Wierling, et al., Systems Biology (Wiley-VCH Verlag GmbH & Co KGaA, Weinheim, Germany, 2009).

A. Kremling, Systems Biology. Mathematical Modeling and Model Analysis (Chapman & Hall/CRC, 2014).

B. Goldstein, J. R. Faeder, and W. S. Hlavacek, Nature Rev. Immunol. 4 , 445 (2004).

A. Friedman and N. Perrimon, Cell 128 , 225 (2007).

A. L. Barabasi and Z. N. Oltvai, Nature Rev. Genet. 5 , 101 (2004).

N. M. Borisov, N. I. Markevich, J. B. Hoek, and B. N. Kholodenko, Biophys. J. 89 , 951 (2005).

C. J. Morton-Firth and D. J. Bray, Theor. Biol. 192 , 117 (1998).

M. L. Blinov, J. R. Faeder, B. Goldstein, and W. S. Hlavacek, Bioinformatics 20 , 3289 (2004).

L. Lok and R. Brent, Nature Biotechnol. 23 , 131 (2005).

M. Rehm, H. J. Huber, H. Dussmann, and J. H. Prehn, EMBO J. 25 , 4338 (2006).

M. M. Hill, C. Adrain, P. J. Duriez, et al., EMBO J. 23 , 2134 (2004).

M. Rehm and J. H. M. Prehn, Methods 61 (2), 165 (2013).

J. Schlessinger, Science 306 , 1506 (2004).

C. J. Marshall, Cell 80 , 179 (1995).

L. O. Murphy, S. Smith, R. H. Chen, et al., Nature Cell Biol. 4 , 556 (2002).

A. Goldbeter and D. E. Koshland, Jr., Proc. Natl. Acad. Sci. U. S. A. 78 , 6840 (1981).

C. Salazar and T. Hofer, J. Mol. Biol. 327 (1), 31 (2003).

N. I. Markevich, J. B. Hoek, and B. N. Kholodenk, J. Cell. Biol. 164 , 353 (2004).

J. E. Ferrell, Jr., Trends Biochem. Sci. 22 , 288 (1997).

W. Xiong and J. E. Ferrell, Jr., Nature 426 , 460 (2003).

A. Goldbeter, Nature 420 , 238 (2002).

J. Stelling, U. Sauer, Z. Szallasi, F. J. Doyle 3rd, and J. Doyle, Cell 118 , 675 (2004).

K. Simons and D. Toomre, Nature Rev. Mol. Cell Biol. 1 , 31 (2000).

A. Sorkin and M. von Zastrow, Nature Rev. Mol. Cell Biol. 3 , 600 (2002).

T. G. Bivona, I. Perez De Castro, I. M. Ahearn, et al., Nature 424 , 694 (2003).

A. Harding, T. Tian, E. Westbury, et al., Curr. Biol. 15 , 869 (2005).

M. Rehm, H. Dussmann, R. U. Janicke, et al., J. Biol. Chem. 277 , 24506 (2002).

N. Fricker, J. Beaudouin, P. Richter, et al., J. Cell Biol. 190 , 377 (2010).

I. N. Lavrik, Cell Death. Dis. 5 , e1259 (2014).

I. N. Lavrik, Curr. Opin. Biotechnol. 21 , 551 (2010).

A. S. Mishin, V. V. Belousov, K. M. Solntsev, and K. A. Lukyanov, Curr. Opin. Chem. Biol. 27 , 1 (2015).

J. Beaudouin, C. Liesche, S. Aschenbrenner, et al., Cell Death. Differ. 20 , 599 (2013).

A. M. Bogdanov, A. S. Mishin, I. V. Yampolsky, et al., Nat. Chem. Biol. 5 , 459 (2009).

L. Neumann, C. Pforr, J. Beaudouin, et al., Mol. Syst. Biol. 6 , 352 (2010).

R. E. Lee, S. R. Walker, K. Savery, et al., Mol. Cell 53 , 867 (2014).

J. G. Albeck, J. M. Burke, S. L. Spencer, et al., PLoS. Biol. 6 , 2831 (2008).

H. Dussmann, D. Kogel, M. Rehm, and J. H. Prehn, J. Biol. Chem. 278 , 12645 (2003)

X. Xia, M. S. Owen, R. E. Lee, and S. Gaudet, Cell Death. Dis. 5 , e1261 (2014).

I. N. Lavrik, A. Golks, and P. H. Krammer, J. Clin. Invest. 115 , 2665 (2005).

S. L. Spencer, S. Gaudet, J. G. Albeck, et al., Nature 459 , 428 (2009).

J. G. Albeck, J. M. Burke, B. B. Aldridge, et al., Mol. Cell 30 , 11 (2008).

S. Pietkiewicz, J. S. Schmidt, and I. N. Lavrik, J. Immunol. Meth. 423 , 99 (2015).

M. Bentele, I. Lavrik, M. Ulrich, et al., J. Cell Biol. 166 , 839 (2004)

Download references

Author information

Authors and affiliations.

Otto-von-Guericke University, Magdeburg, Germany

I. N. Lavrik

Institute of Cytology and Genetics, Siberian Branch, Russian Academy of Sciences, pr. Lavrent’eva 10, Novosibirsk, 630090, Russia

Peter the Great St. Petersburg Polytechnic University, ul. Polytekhnicheskaya 29, St. Petersburg, 195251, Russia

M. G. Samsonova

You can also search for this author in PubMed   Google Scholar

Corresponding author

Correspondence to M. G. Samsonova .

Additional information

Original Russian Text © I.N. Lavrik, M.G. Samsonova, 2016, published in Biofizika, 2016, Vol. 61, No. 1, pp. 93–101.

Rights and permissions

Reprints and permissions

About this article

Lavrik, I.N., Samsonova, M.G. The systems biology of signaling pathways. BIOPHYSICS 61 , 78–84 (2016). https://doi.org/10.1134/S0006350916010127

Download citation

Received : 09 October 2015

Published : 13 May 2016

Issue Date : January 2016

DOI : https://doi.org/10.1134/S0006350916010127

Share this article

Anyone you share the following link with will be able to read this content:

Sorry, a shareable link is not currently available for this article.

Provided by the Springer Nature SharedIt content-sharing initiative

  • signaling pathways
  • mathematical modeling
  • systems biology
  • Find a journal
  • Publish with us
  • Track your research

Pumps & Systems Menu

Subscribe Now and Get:

  • In-depth articles on pump industry issues
  • Expert insights into important topics in the field
  • Keep up with trends, basics, and more

Breadcrumbs

The crucial role of the scientific method & experimental design in engineering innovation.

Enrichments of selenium-reducing microbial consortia identified by the red color of elemental selenium formation in the seepage water matrix

When it comes to wastewater treatment systems, owners demand reliability and effectiveness, but getting from concept to solution is not always a straight path. For this reason, the application of the scientific method 1 (SM) and rigorous experimental design 2 (ED) are essential for advancing engineering designs and ensuring reliable, efficient and innovative engineering solutions.

SM in Engineering

The scientific method can be defined as a process of systematic inquiry using observation, hypothesis, experimentation, analysis and conclusion. Even though most of us are first introduced to SM in grade school, few of us use it in real life, real work and real problem solving on a daily, weekly or even monthly basis.

Even so, having a systematic laboratory investigative mechanism is a cornerstone of engineering practice, as it:

Ensures accuracy with precision and repeatability

Enhances problem solving with a structured approach to identifying and analyzing problems and solutions

Builds credibility and trust through the scientific rigor of engineering studies that can be peer reviewed, critiqued and discussed

Facilitates innovation by allowing scientists and engineers to develop creative and effective solutions to complex problems as well as keys to discovering new methods and technologies 1

ED in Engineering

Experimental design in engineering refers to the structured planning of experiments to investigate and validate hypothesized engineered solutions to users’ wastewater problems. The goal is to gather empirical data that can be analyzed to draw valid and reliable conclusions to make informed decisions. When done correctly, it validates assumptions with experimental data or uncovers unforeseen problems early in the low-cost laboratory study rather than later in the commercial plant, when high-cost modifications are required.

Good ED always states a clear objective that defines the problem to be solved and a hypothetical solution based upon previous experience, literature review and industry knowledge. Good ED includes variables, controls and replicas to statistically validate results. It includes, with constraints where needed, state-of-the-industry approaches to solving wastewater problems.

Enrichments of selenium-reducing microbial consortia identified by the red color of elemental selenium formation in the seepage water matrix

When design of experiment (DOE or ED) is done correctly, it optimizes efficiency and performance of the tested processes and/or products, facilitates development of innovative technologies and solutions and ensures accurate, reproducible results—all while helping to improve the quality of engineered designs and systems. Careful planning leads to meaningful insights and better developed engineered solutions.

Use of SM & ED to Solve a Mine Wastewater Treatment Problem

To better show the role of SM and ED in engineering, here is an example of how these processes were used in the development of a wastewater treatment system for a mine tailings waste stream. In this example, what began as a series of laboratory studies for a mining application ultimately led to the invention of a new selenium-reducing fixed bed bioreactor (Se FBBR) that found use in other applications and the refining industry.

Solution development began in the lab with isolation and observation of anoxic selenium-reducing bacteria (Se-RB) in a mine tailings waste stream (Image 1). Process scientists used SM principles to test hypotheses and better understand the Se-RB culture’s growth dynamics. Based on these experiments, the scientists learned these selenate-reducing bacteria (Se-RB) were also denitrifier bacteria, meaning they were able to grow by reducing nitrate and nitrite in the absence of dissolved oxygen. Additional lab studies were also conducted to determine the best external carbon source to support selenate reduction (a glycerin-based carbon source) as well as the optimal nutrient requirements, temperature and pH range for bacterial growth.

Fixed-bed bioreactor pilot treating selenate-bearing drainage

Through rigorous scientific experimentation, process scientists also discovered that the addition of too much external carbon led to overproduction of hydrogen sulfide, which was an unexpected and undesirable result. This observation enabled the team to innovate a novel aeration/no-aeration sequencing of operational regimes to select and promote growth of the desirable Se-RBs while limiting or killing the growth of unwanted sulfate-reducing bacteria (SRBs). Conceived based on SM and ED principles, these well-designed lab studies allowed the team to investigate potential solutions before testing them in the field, which helped to prevent any surprise findings from cropping up later in the engineering and design process. The analysis from the lab studies led to a full-scale process design and mathematical model, followed by an on-site pilot study (Image 2) to demonstrate and validate the hypothetical model.

A few years later, the selenium FBBR technology developed through this series of lab studies was again put to work, this time for removing selenate from a refinery wastewater stream. The refinery wastewater stream had selenate levels that exceeded limits for discharge into a local surface water stream. Due to high levels of sulfate also present in the wastewater, the selective Se-RBs were needed to out-compete SRBs to preferentially remove selenate while forming minimal hydrogen sulfide.

Challenges & Mitigations

From a water treatment professional’s perspective, the usual challenges to applying the SM and ED in engineering systems include securing the trust of the user to invest in a customized solution-development study, accurately estimating the time and cost of the lab study, obtaining and evaluating the best representative wastewater samples in the lab and designing the sequence of lab experiments to the end solution.

Fabrication of three commercial bioreactors for selenate treatment

Some lab studies are so clear and conclusive that full-scale engineering and warranty can follow, while other studies necessitate an on-site field study to validate the designed approach. In these cases, the field study allows the team to gauge system performance while treating daily plant variations in wastewater concentrations, flow rates, loads and effluent quality. From there, a process warranty can follow. Even if all these phases are needed, a scientifically sound approach is often well worth it in the end, as it can prevent delays and added costs due to unforeseen issues with a solution concept.

Other Examples

SM and ED principles can be leveraged in nearly any type of wastewater treatment system design process. Recent examples of wastewater treatment solutions achieved through application of SM and ED include:

An extensive river remedial action laboratory study involving contaminated sediment dewatering and treatment with subsequent large-scale engineering system design and offer

Conducting a large jar test study to determine the optimal method for removing multiple metals from a highly variable wastewater stream

Conducting a lab study on cyanide and nickel removal from a novel wastewater to support on-site treatment and eliminate expensive disposal costs

Performing a laboratory respirometer study to examine growth kinetics of denitrification bacteria in refinery nitrogen-bearing cooling tower blowdown water

In each of these cases, conducting well-designed laboratory studies at the outset of the project yielded a smooth design and build process thereafter, including the building of a full-scale moving bed bioreactor for treatment of a complex cooling tower blowdown stream. These and other experiences demonstrate the value of using both SM and ED for engagements between wastewater treatment solution providers and users.

Commercial bioreactors installed at site for selenate removal

The use of scientific rigor in wastewater treatment studies with experimental design is extremely important for engineering wastewater treatment solutions and systems. Solution providers should take deliberate efforts to foster a culture of inquiry and experimentation in their approach to wastewater treatment problem solving and solution development.

Adopting SM and ED as standard engineering practice and education will benefit solution providers and clients alike by encouraging earlier detection of problematic issues, more comprehensive problem solving and greater innovation in the engineering process.

Scientific Method, en.wikipedia.org/wiki/scientific method. Elements of inquiry, overview and factors of scientific inquiry.

Introduction to Engineering Experimentation, 3rd ed., 2010. Wheeler & Ganji; sections 12.1.2 through 12.1.6, pp. 423-424.

Further reading: Activated Sludge Technologies for Treating Industrial Wastewaters, Design & Troubleshooting, 2014. Echenfelder & Cleary. Ch 7 section on Treatability Studies (pp. 144-145) and Process Modeling (pp. 163-164).

Bill Sheridan is a senior process design scientist and subject matter expert for Technologies, Inc. He has received four biological wastewater treatment technology patents and has led the licensing of two of those biotechnologies. Additionally, Sheridan has led or conducted close to 100 laboratory treatability studies as well as dozens of field pilot demonstrations and engineering design studies, audits and consultations. He may be reached at [email protected].

Related Articles

  • From Our Partners

Sponsored by

  • Motors & Drives
  • Instrumentation
  • Pump Industry Insider
  • Compressors
  • Browse by Issue
  • Digital Edition
  • Market Analysis
  • Amin Almasi
  • Dominik Fry
  • William Livoti
  • Back to Basics
  • Efficiency Matters
  • Flow Meters
  • Maintenance Minders
  • Sealing Sense
  • HI Pump FAQS
  • Manufacturer's Directory
  • White Papers
  • Sponsored Videos
  • WEFTEC Video Product Showcase - 2023
  • WEFTEC Video Product Showcase - 2022
  • WEFTEC Video Product Showcase - 2021
  • Hydraulic Institute Training Center
  • SWPA Training Center
  • Customer Service
  • Data Privacy

IMAGES

  1. Experimental systems biology approaches reveal interaction mechanisms

    experimental methods in systems biology

  2. Figure 1 from Systems biology: experimental design.

    experimental methods in systems biology

  3. Experimental Methods

    experimental methods in systems biology

  4. Scientific Method

    experimental methods in systems biology

  5. The scientific method is a process for experimentation

    experimental methods in systems biology

  6. A schematic model of system biology for biomarker discovery. Data

    experimental methods in systems biology

VIDEO

  1. How to solve an everyday problem

  2. Exobiology Extant Life Surveyor (EELS) Concept of Operations on Enceladus

  3. Systems Biology

  4. A Brief History of the Society for Experimental Biology (SEB)

  5. Systems Biology 1.1: Differential Equations For Modeling

  6. Unconventional and emerging experimental organisms in cell and developmental biology

COMMENTS

  1. Systems biology primer: the basic methods and approaches

    Abstract. Systems biology is an integrative discipline connecting the molecular components within a single biological scale and also among different scales (e.g. cells, tissues and organ systems) to physiological functions and organismal phenotypes through quantitative reasoning, computational models and high-throughput experimental technologies.

  2. Experimental Methods in Systems Biology

    Learn about the technologies underlying experimentation used in systems biology, with particular focus on RNA sequencing, mass spec-based proteomics, flow/mass cytometry and live-cell imaging. A key driver of the systems biology field is the technology allowing us to delve deeper and wider into how cells respond to experimental perturbations.

  3. Systems biology primer: the basic methods and approaches

    Systems biology is an integrative discipline connecting the molecular components within a single biological scale and also among different scales (e.g. cells, tissues and organ systems) to physiological functions and organismal phenotypes through quantitative reasoning, computational models and high-throughput experimental technologies. Systems biology uses a wide range of quantitative ...

  4. Experimental systems biology approaches reveal interaction mechanisms

    Studies using experimental systems biology approaches start either from a previously identified interaction or from a phenotypic screen testing the hypothesis that interplay exists between microbial species found in communities, and typically focus on a few dominant community members that have been previously identified. ... These methods can ...

  5. Systems biology

    Systems biology is the computational and mathematical analysis and modeling of ... analytic or computational modelling to propose specific testable hypotheses about a biological system, experimental ... methods has become popular among systems biologists to simulate and predict the metabolic phenotypes, using genome-scale models. One ...

  6. Systems biology for beginners

    A web and print Focus on systems biology from Nature Publishing Group provides a practical introduction to a field that for all its promise still has many skeptics. Pick an experimental biologist ...

  7. Course: Experimental Methods in Systems Biology

    About this course: Learn about the technologies underlying experimentation used in systems biology, with particular focus on RNA sequencing, mass spec-based proteomics, flow/mass cytometry and live-cell imaging. A key driver of the systems biology field is the technology allowing us to delve deeper and wider into how cells respond to ...

  8. Advances in systems biology modeling: 10 years of ...

    Advances in experimental methods are generating increasingly larger and complex biological datasets at a faster pace. Often, these data can only be processed and interpreted through the development of quantitative methods, giving birth to what is known as systems biology. ... Systems biology and the understanding of biology through quantitative ...

  9. Experimental Methods in Systems Biology

    This course gives a broad overview of a variety of current experimental techniques used in modern systems biology, with focus on obtaining the quantitative data needed for computational modeling ...

  10. Experimental Methods in Systems Biology

    Experimental Methods in Systems Biology. Start course. Add to calendar. Save for later. Experimental Methods in Systems Biology. Course. en. English. 48 h . This content is rated 0 out of 5-Start course. Source. From www.coursera.org. Conditions. Self-paced. Free Access. Fee-based Certificate. More info. 8 Sequences. Introductive Level.

  11. Systems biology: experimental design

    Learning by experimentation is an iterative process [1]. Prior knowledge about a system based on literature and/or preliminary tests is used for planning. Improve-ment of the knowledge based on first results is followed by the design and execution of new experi-ments, which are used to refine such knowledge (Fig. 1A).

  12. Systems Biology

    About this book. This volume focuses on methodological aspects of systems biology in order to provide a new theoretical approach with a robust and tailored experimental support. Chapters detail mathematical modeling, methodological issues, modulation of the collective behavior, metabolic, dynamic profiling, and quantitative morphological studies.

  13. Experimental Methods in Systems Biology

    A key driver of the systems biology field is the technology allowing us to delve deeper and wider into how cells respond to experimental perturbations. This course gives a broad overview of a variety of current experimental techniques used in modern systems biology, with focus on obtaining the quantitative data needed for computational modeling ...

  14. Experimental design and model reduction in systems biology

    Background In systems biology, the dynamics of biological networks are often modeled with ordinary differential equations (ODEs) that encode interacting components in the systems, resulting in highly complex models. In contrast, the amount of experimentally available data is almost always limited, and insufficient to constrain the parameters. In this situation, parameter estimation is a very ...

  15. Maximizing the Information Content of Experiments in Systems Biology

    Experimental design in systems biology is different from classical experimental design studies. The latter theory was first developed at a time when the number of alternative hypotheses was smaller than the amount of available data and replicates . Systems biology, on the other hand is hypotheses rich and data rarely suffice to decide clearly ...

  16. Experimental Methods in Systems Biology (Coursera)

    Learn about the technologies underlying experimentation used in systems biology, with particular focus on RNA sequencing, mass spec-based proteomics, flow/mass cytometry and live-cell imaging. A key driver of the systems biology field is the technology allowing us to delve deeper and wider into how cells respond to experimental perturbations. This in turns allows us to build more detailed ...

  17. A Bayesian active learning strategy for ...

    Borrowing ideas from Bayesian experimental design and active learning, we propose a new strategy for optimal experimental design in the context of kinetic parameter estimation in systems biology. We describe algorithmic choices that allow to implement this method in a computationally tractable way and make it fully automatic.

  18. Model Selection in Systems Biology Depends on Experimental Design

    Abstract. Experimental design attempts to maximise the information available for modelling tasks. An optimal experiment allows the inferred models or parameters to be chosen with the highest expected degree of confidence. If the true system is faithfully reproduced by one of the models, the merit of this approach is clear - we simply wish to ...

  19. Computational Methods in Systems Biology

    The topics of interest include biological process modelling; biological system model verification, validation, analysis, and simulation; high-performance computational systems biology; model inference from experimental data; multi-scale modeling and analysis methods; computational approaches for synthetic biology; machine learning and data ...

  20. Measuring heritability in messy prebiotic chemical systems

    A key question in origins-of-life research, is whether heritability, and thus evolution, could have preceded genes. While theoretical research has shown that out of equilibrium chemical reaction networks with multiple autocatalytic motifs can provide chemical "memory" and serve as units of heritability, experimental validation is lacking. We established conditions that may be conducive to ...

  21. The systems biology of signaling pathways

    Signaling networks play the central role in the regulation of processes in a single cell and in the entire body. A recent breakthrough in technologies for systems biology, which combine experimental and mathematical methods, permits scientists to model signaling pathways in an individual cell and in cell populations. This approach provides new information on mechanisms that regulate a variety ...

  22. The Crucial Role of the Scientific Method & Experimental Design in

    When it comes to wastewater treatment systems, owners demand reliability and effectiveness, but getting from concept to solution is not always a straight path. For this reason, the application of the scientific method1 (SM) and rigorous experimental design2 (ED) are essential for advancing engineering designs and ensuring reliable, efficient and innovative engineering solutions.