Quantification of parameter uncertainty
Design of probability distributions
In order to create the probability distributions, the location and scale parameters and were required. These can be easily calculated from the mean and standard deviation of the available sample data. However in many cases, there were very little or no reported values for a parameter, or there was a minimum and maximum reported value. It was therefore necessary to come up with an alternative way to derive them which at the same time would be understandable to experimentalists, without demanding complicated mathematical terms and calculations.
In order to achieve this, the mode of the log-normal distribution (global maximum) and its symmetric properties were employed. Log-normal distributions are symmetrical in the sense that values that are times larger than the most likely estimate, are just as plausible as values that are times smaller. More specifically, the mode of the distribution is the value for which the condition for all real numbers , (where is the probability density function) is fulfilled. Hence, the user has to decide on a most plausible value for each parameter, which is set as the mode (global maximum) of the corresponding distribution (Probability Density Function or PDF), and on a range within which lie 95.45% of the values. The latter is linked to the mode via a multiplicative factor, which we call "Confidence Interval Factor". If the mode is multiplied or divided by the CI factor, the range within which 95.45% of the values are found is calculated. For instance, if the most plausible value for a parameter is and the confidence interval multiplicative factor is , then the mode of the distribution is set as the range where 95.45% of the plausible values are found is .
Based on these values, a two-by-two system of the equations containing the cumulative distribution function (CDF) and the mode is solved, in order to derive the location parameter and the scale parameter of the corresponding log-normal distribution. The equations are the following:
where and and are the lower and upper bounds of the confidence interval. By substituting these into the previous equation the final form of the system is obtained:
In this way, the and parameters are obtained and from them it is easy to calculate any property in the distribution (i.e. geometric mean, variance etc.)
Parameter dependency and thermodynamic consistency
In some cases, parameters cannot be chosen separately either because they are statistically dependent, subject to thermodynamic constraints or depend on another common parameter. Therefore, thermodynamic consistency is also an important factor that needs to be considered to decide if the combinations of parameters are plausible. For instance, a very common occurrence in biological systems are forward and backward reactions. The source of dependency is the equilibrium constant, which denotes the relationship between the kinetic parameters for the "on" and "off" components of the reaction. Let’s assume a reaction that is known to have an equilibrium constant very close to 1, i.e. its standard Gibbs free energy � = 0. There is not much information about the rate of the reaction, so each of the two parameters is sampled from a very broad distribution. If the additional thermodynamic information is not taken into account, there will often be cases where values will be sampled from the "fast" end of the spectrum for the forward reaction rate, and from the "slow" end for the backward rate (or vice versa). Thus, inconsistent pairs of the two parameters will be generated. In this case, thermodynamic consistency requires that we discard such samples and only keep those where the two reaction rates are very similar (how similar will in turn depend on our uncertainty about the equilibrium constant).
Additionally, for the parameters that are interconnected (i.e. forward and backward reaction rates) a bivariate distribution was created between , and , in order to account for thermodynamic consistency. As the multivariate system requires a linear dependency between the two marginal distributions, two of the parameters will be independent and the third will be dependent on them. For instance, if the two marginal distributions are and (=), is dependent on the values of and . The parameter with the largest geometric coefficient of variation () is usually set as the dependent one. Any product of two log-normal random variables is also log-normally distributed. Therefore, for the two log-normal distributions and , their product will be the log-normal distribution and its parameters will be , .
A similar strategy applies for the quotient of two log-normal distributions, although in this case the parameter will be derived by the formula . The formula for the calculation of the parameter does not change.
Thus, it becomes easy to transform the two marginal distributions and to normal ones, through the natural logarithm. The problem can therefore be reduced to the case of a multivariate normal distribution generated by the formula
Failed to parse (syntax error): f(x,y)= \frac{1}{2 \pi \sigma_X \sigma_Y \sqrt{1-\rho^2}} \exp\left( -\frac{1}{2(1-\rho^2)}\left[ \frac{(x-\mu_X)^2}{\sigma_X^2} + \frac{(y-\mu_Y)^2}{\sigma_Y^2} - \frac{2\rho(x-\mu_X)(y-\mu_Y)}{\sigma_X \sigma_Y}\right] \right)\\
where is the correlation between and and and . In this case, (covariance matrix).
The required parameter values are obtained by generating samples from the multivariate normal distribution and then exponentiating the results. In order to avoid errors that are introduced to the correlation matrix during the exponentiation, a matlab function called Multivariate Lognormal Simulation with Correlation (MVLOGNRAND) is used, which makes up for these errors.