International
Tables for Crystallography Volume F Crystallography of biological macromolecules Edited by M. G. Rossman and E. Arnold © International Union of Crystallography 2006 |
International Tables for Crystallography (2006). Vol. F. ch. 18.1, pp. 369-370
Section 18.1.4. Least squares and maximum likelihood
a
San Diego Supercomputer Center 0505, University of California at San Diego, 9500 Gilman Drive, La Jolla, CA 92093-0505, USA, and bStructural, Analytical and Medicinal Chemistry, Pharmacia & Upjohn, Inc., Kalamazoo, MI 49001-0119, USA |
`Improving the agreement' between the observed and calculated data can only be done if one first decides the criteria to be used to measure the agreement. The most commonly used measure is the norm of the residuals, which is simply the sum of the squares of the differences between the observed and calculated data (IT C Chapter 8.1
),
where
is the weight of observation
and
is the calculated value of observation i given the parameters x. In essence, least-squares refinement poses the problem as `Given these data, what are the parameters of the model that give the minimum variance of the observations?'. The
norm is strongly affected by the largest deviations, which is not a desirable property in the early stages of refinement where the model may be seriously incomplete. In the early stages, it may be better to refine against the
norm,
the sum of the absolute value of the residuals. At present, this technique is not used in macromolecular crystallography.
The observable quantity in crystallography is the diffracted intensity of radiation. Fourier inversion of the model gives us a complex structure factor. The phase information is normally lost in the formulation of . This is a root cause of some of the problems of least-squares refinement from poor starting models. Many of the problems of least-squares refinement can be addressed by changing the measure of agreement from least squares to maximum likelihood, which evaluates to the likelihood of the observations given the model. In this formulation, the problem is posed as `Given this model, what is the probability that the given set of data would be observed?'. The model is adjusted to maximize the probability of the given observations. This procedure is subtly different from least squares in that it is reasonably straightforward to account for incomplete models and errors in the model in computing the probability of the observations. Maximum-likelihood refinement is particularly useful for incomplete models because it produces residuals that are less biased by the current model than those produced by least squares. Maximum likelihood also provides a rigorous formulation for all forms of error in both the model and the observations, and allows incorporation of additional forms of prior knowledge (like additional phase information) into the probability distributions.
The likelihood of a model given a set of observations is the product of the probabilities of all of the observations given the model. If is the conditional probability distribution of the structure factor
given the model structure factor
, then the likelihood of the model is
This is usually transformed into a more tractable form by taking the logarithm,
Since the logarithm increases monotonically with its argument, the two versions of the equation have maxima at the same values of the parameters of the model. This formulation is described in more detail in Chapter 18.2
, in IT C Section 8.2.1
and by Bricogne (1997
), Pannu & Read (1996
), and Murshudov et al. (1997
).
References


