Accuracy Assessment GD3204
Accuracy Assessment GD3204
Accuracy
Assessment
of
Spa1al
Data
outline
Introduc2on
Spa2al
data
quality
Objec2ve
Type
of
spa2al
data
quality
Posi2onal/Geometric
Accuracy
NSSDA
RMSE
Accuracy
Assessment
Standard
AHribute/Thema2c
Accuracy
Mo2va2on
Error
Matrix
1
3/5/14
2
3/5/14
3
3/5/14
Akurasi
posisi
Positional Accuracy
Reference:U.S. Federal Geographic Data Committee (FGDC)
Geospatial Positioning Accuracy Standards
Part 3: National Standards for Spatial Data Accuracy (NSSDA)
FGDC-STD-007.3-1998
http://www.fgdc.gov/standards/status/sub1_3.html
http://www.fgdc.gov/standards/projects/FGDC-standards-projects/
accuracy/part3/tr96
http://www.mnplan.state.mn.us/pdf/1999/lmic/nssda_o.pdf
4
3/5/14
Positional Accuracy
The NSSDA uses root-mean-square error (RMSE) to estimate
positional accuracy. RMSE is the square root of the average of the
set of squared differences between dataset coordinate values and
coordinate values from an independent source of higher accuracy for
identical points.
Accuracy is reported in ground distances at the 95% confidence level.
This means that 95% of the check points had errors equal to or smaller
than the reported accuracy value.
A minimum of 20 check points shall be tested, distributed to reflect
the geographic area of interest and the distribution of error in the dataset.
When 20 points are tested, the 95% confidence level allows one point
to fall outside the reported accuracy value.
5
3/5/14
See: mcmcweb.er.usgs.gov/sdts/
6
3/5/14
7
3/5/14
RMSerror
Yresidual
TPReference
8
3/5/14
9
3/5/14
Accuracy
Assessment
(standard)
Na2onal
Mapping
Accuracy
Standard
(NMAS),
US.Bureau
of
the
Budget,
1947
uses
level
of
conden2al
90%
(Circular
Mapping
Accuracy
Standard:
CMAS
/
Ver2cal
Mapping
Accuracy
Standard:
VMAS)
FGDC:
Na2onal
Standard
for
Spa2al
Data
Accuracy
(NSSDA)
,
1998
uses
level
of
conden2al
95%
(Circular
Error
:
CE95/
Linear
Error:
LE95)
Linear
Error:
LE
+
LE=RMSE,
LE90%,
LE95%
Circular
Error:
CE
+
CE=RMSE,
CE90%,
CE95%
NSSDA
(Accuracy)
denes
LE95%
and
CE95%
10
3/5/14
11
3/5/14
Worksheet
Accuracy
Assessment
(NSSDA)
12
3/5/14
13
3/5/14
examples:
Minnesota
Department
of
Transporta1on
Map
Scale
1:1000
(from
1:3000
aerial
photography)
Test
to
use:
ver2cal
and
horizontal
accuracy
Data
set:
1.
DEM
(contour
and
grid)
2.
digital
topographic
map
Selec2ng
test
points:
o Ver2cal:
296
random
control
points
(accuracy
10-15mm)
o Horizontal:
40
well-dened
points
(accuracy
10-15mm)
Examples
of
these
include
manholes,
catch
basins
and
right-angle
intersec2ons
of
objects
such
as
sidewalks.
Forty
points
were
chosen
rather
than
the
minimum
of
20
because
they
were
fairly
easy
to
collect
and
because
of
the
long
narrow
shape
of
the
corridor.
Having
the
extra
points
opened
the
possibility
of
comparing
a
test
of
the
20
easternmost
control
points
with
a
test
of
the
20
westernmost
control
points.
14
3/5/14
Akurasi
tema1k
atau
akurasi
atribut
15
3/5/14
Mo2va2on
Classied
thema2c
maps
are
produced
for
a
wide
variety
of
resources:
soil
types
or
proper2es,
land
cover,
land
use,
forest
inventory,
and
many
more.
These
maps
are
not
very
useful
without
quan2ta2ve
statements
about
their
accuracy
Mo2va2on
Map
users
must
know
the
quality
of
the
map
for
their
intended
uses,
and
map
producers
must
evaluate
the
success
of
their
mapping
eorts.
Both
users
and
producers
may
want
to
compare
several
maps
to
see
which
is
best,
or
to
see
how
well
they
agree.
16
3/5/14
Mo2va2on
(Congalton,
Russell
G.)
The
need
for
assessing
the
accuracy
of
a
map
generated
from
any
remotely
sensed
data
has
become
universally
recognized
as
an
integral
project
component.
In
the
last
few
years,
most
projects
have
required
that
a
certain
level
of
accuracy
be
achieved
for
the
project
and
map
to
be
deemed
a
success.
With
the
widespread
applica2on
of
geographic
informa2on
systems
(GIS)
employing
remotely
sensed
data
as
layers,
the
need
for
such
an
assessment
has
become
even
more
cri2cal.
Mo2va2on
There
are
a
number
of
reasons
why
this
assessment
is
so
important,
including
(Congalton,
Russell
G.):
The
need
to
perform
a
self-evalua2on
and
to
learn
from
your
mistakes
The
ability
to
compare
method/algorithms/
analysts
quan2ta2vely
The
desire
to
use
the
resul2ng
maps/spa2al
informa2on
in
some
decision-making
process
17
3/5/14
Mo2va2on
It
is
absolutely
necessary
to
take
some
steps
toward
assessing
the
accuracy
or
validity
of
that
map.
There
are
a
number
of
ways
to
inves2gate
the
accuracy/error
in
spa2al
data
including,
but
not
limited
to:
1. visual
inspec2on
map
look
good,
2. nonsite-specic
analysis,
3. genera2ng
dierence
images,
4. error
budget
analysis,
and
5. quan2ta2ve
accuracy
assessment.
18
3/5/14
Error Matrix
n Inthe evaluation of classification errors, a
classification error matrix is typically
formed.
This matrix is sometimes called confusion
matrix or contingency table.
n In
this table, classification is given as rows
and verification (ground truth) is given as
columns for each sample point.
19
3/5/14
Error Matrix
n The diagonal elements in this matrix indicate numbers of
sample for which the classification results agree with the
reference data.
n Off diagonal elements in each row present the numbers
of sample that has been misclassified by the classifier,
i.e., the classifier is committing a label to those samples which
actually belong to other labels. The misclassification error is
called commission error.
n The off-diagonal elements in each column are those
samples being omitted by the classifier.
Therefore, the misclassification error is also called omission
error.
Error Matrix
20
3/5/14
Error Matrix
n The most common error estimate is the overall
accuracy:
()==1/
n = total number of TPs
Error Matrix
n More specific measures are needed
because the overall accuracy does not
indicate how the accuracy is distributed
across the individual categories.
Thecategories could, and frequently do,
exhibit drastically differing accuracies but
overall accuracy method considers these
categories as having equivalent or similar
accuracies.
21
3/5/14
Error Matrix
n From the confusion matrix, it can be seen that at
least two methods can be used to determine
individual category accuracies.
n (1) The ratio between the number of correctly
classified and the row total
theuser's accuracy - because users are concerned
about what percentage of the classes has been
correctly classified.
n (2) The ratio between the number of correctly
classified and the column total
is called the producer's accuracy.
Error Matrix
n A more appropriate way of presenting the
individual classification accuracies.
Commission error = 1 - user's accuracy
Omission error = 1 - producer's accuracy
==1/
=/+
=/+
+==1
+==1
22
3/5/14
Error Matrix
Error Matrix
23
3/5/14
off .diagonal.row.elements 14 + 15
Comission.Error = = = 0.51
total.of .row 57
off .diagonal.column.elements 1 + 1
Omission.Error = = = .067
total.of .column 30
diagonal. for.class
Mapping.accuracy =
diagonal + off .diag .rows + off .diag .columns
28
= = 0.475
28 + (14 + 15) + (1 + 1)
24
3/5/14
25
3/5/14
Error Matrix
Grand Total = 100, Total correct = 63, Observed correct = 63/100 = 0.63
Pi+ = 0.3 x 0.57 = .171, 0.3 x 0.21 = .063, 0.4 x 0.22 = 0.88
26
3/5/14
Kappa Coefficient
n One of the advantages of using this method is
that we can statistically compare two
classification products.
For example, two classification maps can be made
using different algorithms and we can use the same
reference data to verify them.
Two K s can be derived, K 1, K2. For each K, the
variance can also be calculated.
Another Way
n The following shows an alternative way to
do the error matrix
Errorsof Omission and Commission are both
calculated from the row totals in this
technique
27
3/5/14
28
3/5/14
S
E
L
E
S
A
I
SELAMAT
BELAJAR
29