, 2007, Hatakeyama et al., 2003, Kholodenko et al., 1999 and Klinke, 2010). S.1.4. Definition of the model readouts subject to sensitivity analysis. At this stage
the model readouts for inclusion in the analysis should be specified. In principal, GSA can be applied to any number of model outputs or combination of them, but in practice it is sensible to focus on the analysis of one or several most informative model readouts. For the ErbB2/3 network model we explored the output signal from the PI3K/Akt branch of the network, focusing on the analysis of the time course profile of phosphorylated Akt (pAkt), where pAkt was defined as the composition of several model species, corresponding to different forms of phosphorylated Akt, normalised by the total concentration of Akt protein: pAkt=([pAkt-PIP3]+[ppAkt-PIP3]+[pAkt-PIP3-PP2A]+[ppAkt-PIP3-PP2A])/Akt_totpAkt=([pAkt-PIP3]+[ppAkt-PIP3]+[pAkt-PIP3-PP2A]+[ppAkt-PIP3-PP2A])/Akt_tot this website S.1.5. BLU9931 mw Definition of the criteria to include/reject a
parameter set into/from the analysis. Quasi-random parameter sets sampled from the parameter space correspond to a variety of system behaviours, some of them potentially biologically implausible. Depending on the purpose of the analysis, at this stage the criteria for classifying parameter sets as plausible/implausible should be formulated. For the ErbB2/3 network model, we included in the analysis only those parameter sets, for which the phosphorylation level of Akt in the absence of the drug exceeded 1% of the total Akt protein. Step 2: Sampling N parameter sets from the hypercube To sample the points from the hypercube defined by parameter ranges we use Sobol’s LDS algorithm, which ensures that individual parameter ranges are evenly covered (Joe and Kuo, 2003 and Sobol, 1998), implementation taken from (http://people.sc.fsu.edu/~burkardt/cpp_src/sobol/sobol.html). The choice of the adequate sample size (N) depends on the properties of the system. One way to estimate the optimal N is to systematically increase
the sample size and check, whether the set of the most sensitive parameters keeps changing with the increase of N. When two consecutive experiments consistently capture and rank a similar set through of most important parameters, one can conclude that there is no obvious advantage in further increasing the sample size. For our ErbB2/3 network model we used a quantitative metric “top-down coefficient of concordance” (TDCC) to assess the adequacy of the sample size N, as suggested by Marino et al. (2008). TDCC is a measure of correlation between parameter ranks found in two consecutive sampling experiments, which is designed to be more sensitive to agreement on the top rankings ( Iman and Conover, 1987). We calculated TDCC for sample size N = [5000, 10,000, 30,000, 40,000, 50,000, 80,000, 100,000, 120,000].