44
servation data are treated here.
Initially omitting prior information on parameters from the regression encourages under-
standing of the information directly available from the observations. Two
reasons generally would
motivate the use of prior information. First, if the sensitivity for a parameter is low, as indicated by
a small composite scaled sensitivity, regression including the parameter often will not converge.
Two possibilities generally exist: specify prior information on the parameter or set the parameter
value so that it is not changed during the regression (which is roughly equivalent to prior informa-
tion with a very large weight). Specifying prior information usually will result in a
parameter esti-
mate that is close to the value specified in the prior information, so that the estimate will be equal
or close to the prior value regardless of which option is chosen. Execution time is less when the
parameter value is set because this eliminates the need to calculate sensitivities for the parameter,
so it is suggested that this option be followed for model calibration. This will continue to be the
best option as long as the parameter remains insensitive, which can be
checked during calibration
by occasionally calculating composite scaled sensitivities for the estimated parameters and the pa-
rameter in question. An exception to this guideline occurs when the user purposely defines more
parameters than can be directly supported by the data to represent suspected system complexity,
and this generally requires substantial use of prior information to obtain a well-posed regression
problem. An example of this use of prior information and its effect on model accuracy is presented
in a synthetic test case by Hill and others (1998).
The other common reason for using prior information on parameters
is when the parameter
value estimated by the regression is unreasonable. This problem is discussed in the previous sec-
tion of this report titled "Lack of Limits on Estimated Parameter Values." As noted there, the most
productive response to this problem depends on the amount of information the observations pro-
vide on the parameter in question. If little information is provided, the problem falls into the cate-
gory
of insensitive parameters, and the guidelines discussed in the paragraph above apply. If
substantial information is provided, the unrealistic estimated parameter value is likely to indicate
problems with the model or the data, as discussed by Anderman and others (1996) and Poeter and
Hill (1996). To determine whether enough information is provided by the observations such that
the unrealistic estimated parameter value indicates a problem with the model
or the observations,
the linear confidence interval on the parameter can be considered. If the confidence interval in-
cludes no realistic parameter values, the unrealistic estimate is likely to indicate problems with the
model or the observations. If the confidence interval includes realistic parameter values, it is not
clear whether there is a problem with the model or the data. Examples of
the first circumstance are
described by Anderman and others (1996), Poeter and Hill (1996), and Hill and others (1998). An
example of the latter circumstance is described by Christiansen and others (1995) and Barlebo and
others (in press) for a problem in which only hydraulic-head observations are used. In that appli-
cation, addition of concentration observations produced more realistic parameter values, indicating
that the problem was primarily due to inadequate data. UCODE and MODFLOWP prints linear