Professional Documents
Culture Documents
*Density of Observations
-an insufficient number of observations may
not provide the required level of resolution
-ex. If you have a 40’ contour interval, you
had better not be reporting on or making
decisions about features only a few feet in
difference
Obvious Sources of Error
*Relevance
-surrogate data may be used to indirectly
describe/classify/quantify features
-Ex. We can create a forest polygon layer from
classification of remotely sensed data.
However, we are not classifying a “tree” as a
tree. Rather, we are classifying the imagery
based on spectral signatures, and those
signatures can be related to tree species.
Obvious Sources of Error
*Format
-methods of formatting data can introduce
errors
-conversion of scale, projection, or datum,
vectorization/rasterization, and pixel
resolution are possible areas of format error
-international mapping standards not
established
Obvious Sources of Error
*Accessibility
-try getting a highway map of the former
USSR in the Cold War days . . . Good Luck!
*Cost
-highly accurate, precise data is expensive!!!
Errors from Natural Variation
or from Original Measurements
*Positional Accuracy
-many natural features do not exhibit “hard”
boundaries like roads or boundary lines
-examples include . . .?
Errors from Natural Variation
or from Original Measurements
*Positional Accuracy
-many natural features do not exhibit “hard”
boundaries like roads or boundary lines
-examples include:
-soils
-vegetation communities
-climate variables
-drainage
-biomes, etc.
Errors from Natural Variation
or from Original Measurements
*Accuracy of Content
-qualitative accuracy refers to correct
labelling/classification (Ex.-pine forest vs.
spruce forest)
-quantitative inaccuracies often occur from
faulty equipment or poor readings
-what forestry equipment could give you bad
data? And how?
Errors Arising Through Processing
*Numerical Errors
-by far, the hardest errors to detect!!!
-different (faulty) computer chips can compute
differently, generating a different output
(response)
*Topological Errors
-overlaying, or deriving/creating new variables
based on other data can cause slivers,
overshoots, and dangles
Errors Arising Through Processing
*Classification/Generalization Errors
-classification inaccuracies/class merging
-grouping data in different ways can lead to
dramatically different results (Ex.-studying
cause of death amongst males would probably
be quite different if you had (amongst others)
an aged 18-25 group vs. an 18-50 group
Errors Arising Through Processing
*Geocoding/Digitizing Errors
-what can cause digitizing errors?
Errors Arising Through Processing
*Geocoding/Digitizing Errors
-what can cause digitizing errors?
-rasterizing will cause positional error
Error, Error, Everywhere . . .
How can we manage error?
1. Be aware of where error can be
generated (everything discussed in this
presentation)
2. Metadata, metadata, metadata . . . Fully
understand all data compiled for your GIS,
make notes of all work done with the data,
and send such information to future users
or with all GIS generated output.