• Ei tuloksia

PART III: DEVELOPMENT AND EVALUATION OF MODEL IDENTIFICATION

9.   PARAMETER IDENTIFICATION

9.2   U NCERTAINTIES OF MRF M ODEL P ARAMETER E STIMATES

The MPL method yields only optimal parameter values but ignores the uncertainties related to these parameter estimates. In pseudolikelihood approximation, parameter uncertainties can be assessed by the Bayesian approach, discussed in Chapter 6. Because of no prior information about MRF model parameters, the prior probability distribution in Bayes’s theorem can be set as a constant and thus, according to Eq. (6.7), with pseudolikelihood approximation, the posterior distribution approximately equals the product of the constant multiplying the pseudolikelihood

term: | ∏ ∏ | , . When conditional independence

prop-erties are further used, as in Eq. (9.5), the pseudolikelihood posterior parameter distribution be-comes now

This is the general form of the parameter probability distribution of an MRF model, when prior information is assumed absent and the pseudolikelihood approximation is given by Eq. (6.6). The distribution reveals all the available information about the uncertainties of the MPL parameter estimates. Because MPL parameter estimates MPL are obtained by maximising Eq. (9.7), the mode of the posterior distribution is obviously MPL. In the general MRF model, the posterior distribution | is not known in its analytical form, but in some special cases it is. For example, in a GMRF model, all conditional distributions are Gaussians, thus also the posterior distribution. In the general case, uncertainty is not as easily assessed as in the Gaussian case.

However, the general posterior distribution can be approximated with a Gaussian distribution, and if the approximation is appropriate, the uncertainty of the Gaussian distribution, i.e., that given by the covariance matrix, can be considered a parameter uncertainty of the posterior distri-bution. To obtain a Gaussian approximation, let us first denote the gradient (first-order deriva-tive) and Hessian (second-order derivaderiva-tive) operators with respect to parameters by and , and make a second-order Taylor series approximation to the logarithm of the posterior distribu-tion | at around MPL as

The gradient term goes to zero, because the estimate maximises the posterior distribution. By taking the exponent from both sides and denoting | | MPL and

log | | MPL (inverse covariance matrix), Eq. (9.8) is simplified to

This is the Gaussian approximation of the true parameter posterior distribution | at

MPL with a covariance matrix , or precision matrix . If the Gaussian approximation is close enough to the true posterior distribution, uncertainties in the estimated model parameters

| | , . (9.7)

can now be studied through the covariance matrix , where the diagonal elements give the vari-ances of the respective parameters and the off-diagonal elements inform us about the correla-tions between the parameter uncertainties.

By denoting the log-pseudolikelihood by , log | , the inverse covariance matrix is obtained for the Ising model parameters , , as

whose matrix elements are specified in Table 9.1.

To verify the fitness of the Gaussian approximation, the approximate Gaussian distribution must be compared to the respective true posterior distribution, which can be done as follows. First, the true posterior distribution | is evaluated in a region of parameter space where probability values are (clearly) non-zero. Approximately this region can be found as an ellipsoid of the Gaussian contour of the constant probability of the approximative Gaussian distribu-tion. Parameter values can then be sampled from the area of the ellipsoid and the true posterior distribution evaluated at these points. The true posterior distribution and the Gaussian

approxi-, MPL

, , ,

, , ,

, , ,

MPL

, (9.10)

Table 9.1. Elements of defined in Eq. (9.10) for Ising model parameters , , .

, s cosh ,

, cosh ,

, cosh ,

, , cosh , s

, , cosh , s

, , tanh , cosh ,

mation can now be compared by evaluating the marginal distribution for each parameter for both true and Gaussian approximation of the posterior distribution, and by then comparing the re-spective marginal distributions obtained. If the marginal distributions are similar, the approxima-tion is sound and can be used to study parameter uncertainties.

Let us further examine the practical calculation of true marginal probability distributions in the Ising model, and assume that a set of parameter values, , , , is uniformly and ran-domly drawn inside the contour of the constant probability with the superscript indexing the sample values 1, … , . Given an observed data set, , , 1, … , , the mar-ginal distribution of each parameter is obtained similarly. Here is considered an example for which the marginal distribution is derived by marginalising the joint distribution as

where denotes the domains of parameters and . However, because the data set of pa-rameter values, , , , is sampled, the true marginal distribution is obtained in practice by first dividing the sample values of , , into a set of constant-size intervals, say

Δ , with 1, … , indexing the intervals. Then Eq. (9.11) is applied to each interval separately, and at a certain interval Δ , the marginal probability of parameter is obtained by summing over all probability values of those sample values that appear at that interval, and then normalising the values to probabilities:

Here Δ is the width of the interval Δ , and considered a constant. The denominator in Eq. (9.12) normalises the distribution to probabilities. Hence values form a dis-crete probability distribution estimate of the true marginal probabilities of inside the contour of the constant probability .

According to the definition of pseudolikelihood in Eq. (6.7), the joint probability distribution estimate of the parameters can be written by using the pseudolikelihood, , as , ,

. Writing Eq. (9.12) by using this formula and the marginalisation in Eq. (9.11) yields

where the terms appearing both in the nominator and the denominator cancels. In practice, parameters are estimated from the log-pseudolikelihood , log  , which is computa-tionally easier to evaluate than evaluating the pseudolikelihood directly. Hence when the log-pseudolikelihood is used in Eq. (9.13), it becomes

| |

In some practical calculations, taking the exponential in Eq. (9.14) from , may not be suc-cessful, because its values may be very large. It may then help to add a constant to the log-pseudolikelihood values inside the exponents of both the nominator and denominator in Eq.

(9.14). These values cancel out each other, and if the constant is chosen appropriately, the log-pseudolikelihood values can be reduced to enable practical calculation.

Eq. (9.14) can be directly used to estimate the true marginal probabilities of parameter , when log-pseudolikelihoods have been calculated for sampled parameter values. Marginal probability is calculated similarly for the other two parameters. In the Gaussian approximation, a marginal probability distribution is obtained for each parameter by simply evaluating a univariate Gaussian distribution, with the expectation value determined by the respective maximum pseudolikelihood parameter value and variance by the respective diagonal element in the estimated covariance ma-trix .