The effects of five data processing methods on the accuracy of
geostrophic computations are analyzed using hydrographic data collected
from a set of 24 hour triangular stations off the Oregon coast.
Sources of error are assessed by comparison with an available in
situ Salinity/Temperature/Depth record (STD) trace at a control
The results of this study show that the accuracy of the dynamic
depth anomaly depends on the choice of vertical integration intervals
based on sea water density distribution instead of on an arbitrary set
of standard depths for all the cases and that the effect of interpolation
of hydro- data is not significant if the data is properly sampled
and adequately interpolated.
Some sources of error in computations of geostrophic currents
from hydrographic data can be alleviated by proper sampling and
processing procedures. They are (1) the sampling errors caused by
improper sampling schemes which miss the sharp features and
periodic variations (i. e. internal waves) in the ocean, and (2) processing
errors caused by improper interpolations of data and determinations
of mean density.