Академический Документы
Профессиональный Документы
Культура Документы
MAHESH PAL
NATIONAL INSTITUTE OF TECHNOLOGY
KURUKSHETRA, INDIA
Spectral range
(microns)
0.450 - 0.515
0.525 - 0.605
0.630 - 0.690
0.750 - 0.900
1.550 - 1.750
10.40 - 12.50
2.090 - 2.350
0.520 - 0.900
Ground
resolution (m)
30
30
30
30
30
60
30
15
Bands (79)
Wavelength range
(micrometer)
VIS/NIR
32
0.50 - 1.05
SWIR I
1.50 - 1.80
SWIR II
32
1.90 - 2.50
MIR
3.00 - 5.00
TIR
8.70 - 12.50
Importance of a Hyperspectral
Sensor
CLASSIFICATION
Land cover classification has been a major research
area involving the use of remote sensing images.
Image classification process involves in assigning
pixels to classes in terms of the characteristics of
the objects or materials.
A major input in GIS based studies
Several approaches are used for land cover
classification.
CLASSIFICATION ALGORITHMS
Predictive accuracy
Computational cost
o
o
Robustness
o
Interpretability:
o
Landgrebe, 1994)
3. Use of some form of dimensionality reduction procedure prior to
the classification analysis.
Training samples
Learning algorithm
Model/ function
Also called as
Hypothesis
Output values
Testing samples
Hypothesis can be considered as a machine that provides the prediction for
test data
Empirical
risk
SVM
Map data from the original input feature
space to a very high dimensional feature
space.
Data becomes linearly separable but
problem becomes computationally difficult.
Kernel function allows SVM to work in
feature space, without knowing mapping
and dimensionality of feature space.
Advantages
Margin
theory
suggest
dimensionality of input space
no
effect
of
90
85
80
75
70
65
60
8 pixels
15 pixels
25 pixels
50 pixels
75 pixels
100 pixels
55
5
10
15
20
25
30
35
40
45
50
55
60
65
Number of features
Mahesh Pal and Giles M. Foody, 2010, Feature selection for classification of hyperspectral data
by SVM. IEEE Transactions on Geoscience and Remote Sensing, Vol. 48, No. 5, 2297-2306.
Disadvantages
MAHESH PAL AND G.M FOODY, 2012, Evaluation of SVM, RVM and SMLR for accurate image classification
with limited ground data, IEEE journal of selected topics in applied earth observations and remote sensing, 5( 5),
1344-1355.
support
vectors
represent
the
least
prototypical
examples
(closer
to
boundary, difficult to
classify),
relevant
vectors are the most
prototypical
(more
representative of class)
Disadvantages
Requires
SVM.
Choice of kernel
May have a problem of local minima
Band 5
90
80
Wheat
70
Sugar beet
Oilseed rape
60
50
40
70
80
90
Band 1
100
PRESENT WORK
Working
with
COST
Action
(European
Cooperation in Science and Technology)
TD1202: Mapping and the citizen sensor as Non
EU member
1. Classification with imperfect/noisy data
2. How SVM / RVM and SMLR works with noisy
data
3. Will be working on other classifiers- RF, ELM
Error in
data -
5%
RVM
88%
(51)
SMLR
SVM
10%
15%
20%
25%
30%
35%
40%
84.22% 83.11%
(447)
(490)
Multiclass
Very fast
Classification Accuracy
Dataset
SVM (%)
KELM (%)
ETM+
88.37
90.33
ATM
92.50
94.06
DAIS
91.97
92.16
Computational cost
Dataset
SVM (sec)
KELM (sec)
ETM+
76.74
5.78
DAIS
40.78
1.02
ATM
1.30
0.17
Mahesh Pal, A.E. Maxwell and T. A. Warner, Kernel based Extreme Learning Machine for Remote Sensing Image
Classification,2014, Remote Sensing letters.
PRESENT WORK
Working
on
sparse
extreme
learning
FEATURE REDUCTION
Feature
reduction
may
speed-up
the
classification process by reducing data set size.
FEATURE EXTRACTION
Number of techniques for feature extraction including
Principal Components,
maximum noise fraction
transformation, non-orthogonal techniques such as
projection pursuit, Independent component analysis are
proposed.
MNF requires estimates of the signal and noise
covariance matrices
Different features provided by MNF are ranked as per
signal-to-noise ratio (First MNF have smallest value of SN ratio).
Results with DAIS data suggests that MNF may not be
used effectively for dimensionality reduction.
Feature selection
Three approaches of feature selection are:
Filters: uses a search algorithm to search through the space of
possible features and evaluate each feature by using a filter such as
correlation and mutual information
Wrappers: uses a search algorithm to search through the space of
possible features and evaluate each subset by using a classification
algorithm.
Embedded: some classification processes such as random forest/
Multinomial logisitic regression produce a ranked list of features
during classification.
Filters
Large number of filter based approach are available in literature.
Some used with hyperspectral data are:
1. Correlation-based feature selection
2. Minimum-Redundancy-Maximum-Relevance (mRMR)
3. Entropy
4. Fuzzy entropy
5. Signal-to-noise ratio
6. RELIEF
WRAPPER APPROACH
SVM-RFE approach utilise SVM as base classifier.
1 2 w
The SVM-RFE utilise the objective
function
as a feature ranking criterion to produce a list of
features ordered by their discriminatory ability.
The feature, with the smallest ranking score is
eliminated.
SVM-RFE uses a backward feature elimination scheme
to recursively remove insignificant features from subsets
of features in order to derive a list of all features in rank
order of value.
A major drawback of wrapper methods is their high
computational requirements
EMBEDDED APPROACH
During classification process some algorithm produce
ranked list of all features.
For example: two approaches based on Random forest
and Multinomial logistic regression classifier can be
used.
In contrast to the filter and wrapper approaches, the
search for an optimal feature subset by embedded
approach is built into the classification algorithm
itself.
Classification and the feature selection process
cannot be separated.
Data Set
1. DAIS 7915 sensor by German Space Agency flown on 29 June
2000.
2. The sensor acquire information in 79-bands at a spatial resolution of
5m in the wavelength range of 0.50212.278 m.
3. 7 features located in the mid- and thermal infrared region and 7
features from spectral region of 0.502 2.395 m due to striping
noise were removed.
4. An area of 512 pixels by 512 pixels and 65 features covering the test
site was used.
Feature selection
Algorithm
Number of used
features
Accuracy
65
91.76
Fuzzy entropy
14
91.68
Entropy
17
91.61
20
91.68
Relief
SVM-RFE
20
88.61
None
mRMR
CFS
Random forest
Multinomial logistic
regression
13
91.89
37
17
21
91.84
91.84
92.08
15
92.76
PRESENT WORK