Вы находитесь на странице: 1из 8

International Journal of Application or Innovation in Engineering & Management (IJAIEM)

Web Site: www.ijaiem.org Email: editor@ijaiem.org


Volume 9, Issue 3, March 2020 ISSN 2319 - 4847

Remote Sensing Image Category Classification


Using Deep Learning
Dr L.JabaSheela, R.Banupriya, G.Dilisha and J.Princy
Professor, Department of Computer Science and Engineering, Panimalar Engineering College, TamilNadu , India.

Student, Department of Computer Science and Engineering, Panimalar Engineering College, TamilNadu, India.

Student, Department of Computer Science and Engineering, Panimalar Engineering College, TamilNadu, India.

Student, Department of Computer Science and Engineering, Panimalar Engineering College, TamilNadu, India.

ABSTRACT
The remote sensing Image Classification plays a major role in real-time applications. Deep Learning plays a vital
role in different fields such as Natural language Processing, Computer Vision medical fields, and Image
classification. Compared to the machine learning algorithms, deep networks provide higher accuracy, strong ability
to learn data extraction. Geographical satellite images that are utilized for the investigation of environmental and
geological fields are acquired through remote sensing techniques. The rough pictures accumulated from the
satellites are not suitable for authentic assessment and exact report plans, so crude pictures experience the
customary picture grouping systems, for example, information preprocessing, division, information include
extraction and characterization. The old picture characterization techniques have spatial and otherworldly goals
issues. The most recent picture order strategy specifically profound CNN systems. The CNN algorithm classifies the
images into various categories namely water, land, forest, agricultural area.
Keywords: Image Classification, Enhancement, Remote Sensing, Resolution, Satellite Sensors, deep learning,
Convolution neural networks.

1. INTRODUCTION
The Earth Observation is a strategy of gathering information about planet Earth through remote detecting. The region,
where we can accumulate the most data about our planet, in the case of evaluating, cultivation, catastrophic event, oil
subordinate and minerals unmistakable confirmation, mapping of the land use, and so on. The eminent body satellites
make the first-rate photos of the entire earth in a less proportion of the time. The photos conveyed by the land satellites
have a great deal of upheaval and unessential data due to the interferences caused in the space. Real-time information
gathered from Deepsat SAT-4. Remote Sensing information collection, Remote detecting information preprocessing,
Model Training and Testing, Image Classification utilizing Deep CNN are the modules used to arrange the remote
detecting satellite pictures.

2. RELATED WORK
Remi Ratajczak , et.al [1] stated “The land spread recreation from monochromatic chronicled ethereal pictures is a
difficult assignment that has as of late pulled in an expanding enthusiasm from mainstream researchers with the
expansion of huge scale epidemiological examinations including review investigation of spatial examples.
Notwithstanding, the endeavors made by the PC vision network in remote-detecting applications are for the most part
centered around forthcoming methodologies through the investigation of high-goals multi-phantom information
procured by the progressed spatial projects. Subsequently, four commitments are proposed right now. They target
giving an examination premise to the future advancement of PC vision calculations applied to the computerization of
the land spread remaking from monochromatic verifiable elevated pictures. Initial, another multi-scale multi-date
dataset made out of 4.9 million non-covering explained patches of the France domain somewhere in the range of 1970
and 1990 has been made with the assistance of topography specialists. This dataset has been named HistAerial. Second,
a broad examination investigation of the cutting edge surface highlights extraction and order calculations, including
profound convolutional neural systems (DCNNs), has been performed. It is exhibited as an assessment. Third, a novel
low-dimensional neighborhood surface channel named pivoted corner nearby twofold example (R-CRLBP) is exhibited
Volume 9, Issue 3, March 2020 Page 66
International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org
Volume 9, Issue 3, March 2020 ISSN 2319 - 4847

as a rearrangement of the double slope shapes channel using asymmetrical mix portrayal. At last, a novel mix of low-
dimensional surface descriptors, including the R-CRLBP channel, is presented as a light blend of neighborhood parallel
examples (LCoLBPs). The LCoLBP channel accomplished best in class results on the HistAerial dataset while saving a
moderately low-dimensional element vector space contrasted and the DCNN approaches”.

Gang Zheng JianguoLiu, et.al[2] stated “A novel method for naturally finding tropical tornado (TC) fixates dependent
on top cloud movements in sequential geostationary satellite pictures. The high imaging rate and spatial goals pictures
of the Gaofen-4 geostationary satellite empower us to infer pixel-wise top cloud movement information of TCs, and
from the information, TC winding focuses can be precisely decided dependent on a totally extraordinary guideline from
those dependent on static picture highlights. Initial, a physical movement field disintegration is proposed to kill scene
move and TC relocation in the movement information without requiring any assistant geolocation information.
This disintegration doesn't create the ancient rarities that show up in the aftereffects of the recently distributed
movement field decay. At that point, a calculation of a movement heading based list inserted in a pyramid looking
through the structure is completely intended to naturally and successfully find the TC communities. The test shows that
the TC concentric movements are all the more unmistakably uncovered after the proposed movement field decay and
the found places are in acceptable concurrence with the cloud design focuses in a visual sense and furthermore with the
best track informational indexes of four meteorological organizations”.

SinaGhassemi, et.al [3] stated “The problem of training a deep neural network for satellite image segmentation so that
it can be deployed over images whose statistics differ from those used for training. For example, in post-disaster
damage assessment, the tight time constraints make it impractical to train a network from scratch for each image to be
segmented. We propose a convolutional encoder-decoder network able to learn visual representations of increasing
semantic level as its depth increases, allowing it to generalize over a wider range of satellite images. Then, we propose
two additional methods to improve network performance over each specific image to be segmented. First, we observe
that updating the batch normalization layers’ statistics over the target image improves the network performance
without human intervention. Second, we show that refining a trained network over a few samples of the image boosts
the network performance with minimal human intervention. We evaluate our architecture over three data sets of
satellite images, showing the state-of-the-art performance in binary segmentation of previously unseen images and
competitive performance with respect to more complex techniques in a multi-class segmentation task.
Satellite image segmentation has received lots of attention lately due to the availability of annotated high-resolution
image data sets captured by the last generation of satellites. The problem of segmenting a satellite image can be defined
as classifying (or labeling) each pixel of the image according to a number of classes, such as buildings, roads, water,
and so on (semantic pixel labeling). Recent research in semantic pixel labeling builds upon and leverages recent
advances in supervised image classification achieved with convolutional neural networks (CNNs). CNN's are artificial
feedforward, acyclic, neural networks typically composed of a feature extraction stage followed by a classification stage.
This paper tackles the challenging case where the segmentation algorithm is to be deployed over images that are not
known at training time. Indeed, most CNN-based schemes for satellite image segmentation focus on the case where the
network is trained over images similar to those where the algorithm is to be deployed. However, if training and test
images are captured by different sensors or at different time intervals or locations, they exhibit different statistics, an
issue is known as covariate shift. In some applications, such as emergency mapping, satellite images must be
segmented in a short time in the aftermath of events, such as floods or earthquakes. In similar scenarios, the tight time
constraints prompt solutions that allow reusing some algorithms previously trained over different images.
Satellite picture division has gotten heaps of consideration of late because of the accessibility of commented on high-
goals picture informational collections caught by the last age of satellites. The issue of fragmenting a satellite picture
can be characterized as arranging (or naming) every pixel of the picture as indicated by various classes, for example,
structures, streets, water, etc (semantic pixel marking). Ongoing examination in semantic pixel marking expands upon
and use late advances in administered picture order accomplished with convolutional neural systems (CNNs) . CNN's
are fake feedforward, non-cyclic, neural systems commonly made out of an element extraction arrange followed by an
order organize.
This paper handles the difficult situation where the division calculation is to be conveyed over pictures that are not
known at preparing time. For sure, most CNN-based plans for satellite picture division center around the situation
where the system is prepared over pictures like those where the calculation is to be conveyed. Notwithstanding, if
preparing and test pictures are caught by various sensors or at various time interims or areas, they show various
measurements, an issue known as covariate move. In certain applications, for example, crisis mapping, satellite
pictures must be divided in a brief time frame in the consequence of occasions, for example, flood or quake. In

Volume 9, Issue 3, March 2020 Page 67


International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org
Volume 9, Issue 3, March 2020 ISSN 2319 - 4847

comparative situations, the tight time limitations brief for arrangements that permit reusing a few calculations recently
prepared over various pictures”.

Bo Du, et.al[4] stated “Item following is a hotly debated issue in PC vision. On account of the blasting of the extremely
high goals (VHR) remote detecting methods, it is presently conceivable to follow focuses of interests in satellite
recordings. Be that as it may, since the objectives in the satellite recordings are typically excessively little in correlation
with the whole picture, and excessively comparable with the foundation, best in class calculations neglected to follow
the objective in satellite recordings with acceptable precision. Because of the way that the optical stream demonstrates
the extraordinary potential to distinguish even the slight development of the objectives, we proposed a multi-frame
optical stream tracker for object following in satellite recordings. The Lucas–Kanade optical stream technique was
intertwined with the HSV shading framework and necessary picture to follow the objectives in the satellite recordings,
while a multi-frame distinction strategy was used in the optical stream tracker for a superior translation. The analyses
with five VHR remote detecting satellite video datasets demonstrate that contrasted and cutting edge objects following
calculations, the proposed strategy can follow the objective all the more precisely”.

Kia Zhang, et.al[5] stated “Inadequate coding-based picture combination techniques have been grown widely. Albeit a
large portion of them can create serious combination results, three issues should be tended to 1) these techniques isolate
the picture into covered fixes and procedure them autonomously, which disregard the consistency of pixels in covered
patches; 2) the segment system brings about the loss of spatial structures for the whole picture, and 3) the connection in
the groups of multispectral (MS) picture is overlooked. Right now, propose a novel picture combination strategy
dependent on convolution structure meager coding (CSSC) to manage these issues. Initially, the proposed strategy
consolidates convolution inadequate coding with the debasement relationship of MS and panchromatic (PAN) pictures
to build up a reclamation model. At that point, CSSC is explained to portray the connection in the MS groups by
presenting auxiliary sparsity. At long last, include maps over the built high-spatial-goals (HR) and low-spatial-goals
(LR) channels are registered by elective advancement to recreate the melded pictures. In addition, a joint HR/LR
channel learning structure is likewise portrayed in detail to guarantee consistency and similarity of HR/LR channels.
Attributable to the immediate convolution on the whole picture, the proposed CSSC combination strategy evades the
parcel of the picture, which can proficiently misuse the worldwide relationship and protect the spatial structures in the
picture. The exploratory outcomes on Quick Bird and GeoEye-1 satellite pictures show that the proposed technique can
deliver better outcomes by visual and numerical assessment when contrasted and a few notable combination strategies”.

3. PROPOSED SYSTEM
As of late, there has been an expanding interest for applications to screen the objectives identified with land-use,
utilizing remote detecting pictures. Proposed the programmed way to deal with restricting and distinguish building
impressions, street systems, and vegetation zones. A programmed understanding of visual information is a thorough
errand in the PC vision field. Deep learning approaches improve the capacity of order in a keen manner. Profound
Learning calculations give high exactness contrasted with the semi-regulated AI calculations. Our Proposed
calculations give High Speed on Testing and training data. It is material on Hyper ghastly and VHR remote detecting
pictures, for example, building impressions, street systems, and vegetation territories. Programmed understanding of
visuals.

MODULES:
• Remote Sensing data collection
• Remote sensing data preprocessing
• Model Training and Testing
• Image Classification using Deep CNN

REMOTE SENSING DATA COLLECTION:


Real-time data collected from Deepsat SAT-4. The collection of data is one of the major and most important tasks of
any machine learning projects. Because the input we feed to the algorithms is data. So, the efficiency and accuracy of
the algorithm depend upon the correctness and quality of data collected. So as the data same will be the output.

Volume 9, Issue 3, March 2020 Page 68


International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org
Volume 9, Issue 3, March 2020 ISSN 2319 - 4847

REMOTE SENSING DATA PREPROCESSING:


Collecting the data is one task and making that data useful is an an-other vital task. Data collected from various means
will be in an unorganized format and there may be a lot of null values, invalid data values, and unwanted data.
Cleaning all these data and replacing them with appropriate or approximate data and removing null and missing data
and replacing them with some fixed alternate values are the basic steps in the pre-processing of data. Even data
collected may contain completely garbage values. It may not be in the exact format or way that is meant to be. All such
cases must be verified and replaced with alternate values to make data meaning meaningful and useful for further
processing. Data must be kept in an organized format.

MODEL TRAINING AND TESTING:


Finally, after the processing of data and training is, the very next task is testing. This is where the performance of the
algorithm, quality of data, and the required output all appear out. From the huge data set collected 80 percent of the
data is utilized for training and 20 percent of the data is reserved for testing. Training as discussed before is the process
of making the machine learn and giving it the capability to make further predictions based on the training it took.
Whereas testing means already having a predefined data set with output also previously labeled and the model is tested
whether it is working properly or not and is giving the right prediction or not. If the maximum number of predictions is
right then the model will have a good accuracy percentage and is reliable to continue with otherwise better to change
the model.

IMAGE CLASSIFICATION USING DEEP CNN:


The next step is algorithms that are applied to data and results are noted and observed. Deep CNN applied to improve
accuracy at each stage.

4. SYSTEM ARCHITECTURE

5. APPLICATIONS
The main applications of remote sensing include Agriculture, Forestry, Geology, Hydrology, Sea ice, land cover
Mapping, Oceans, and Coastal.

5.1 AGRICULTURE:
To look at the soundness of yields, airborne pictures are utilized. It additionally incorporates observing cultivating
rehearses

Volume 9, Issue 3, March 2020 Page 69


International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org
Volume 9, Issue 3, March 2020 ISSN 2319 - 4847

Figure 1 Classification of agricultural field


5.2 FORESTRY:
Ranger service uses of remote detecting incorporate ecological observing, business ranger service and study mapping
To meet the destinations set by national timberland and natural divisions, remote detecting is utilized. It incorporates
refreshing woods spread, estimating biophysical qualities. Business ranger service applications, for example, checking
vegetation thickness and estimating biomass parameters. It additionally incorporates checking woods wellbeing,
amount and decent variety.

Figure 2 Classification of forest area

5.3 GEOLOGY:
Remote detecting is an essential device for mapping land highlights, for example, Structural mapping, lithological
mapping, and rock mapping. It is additionally used to extricate the data about the land surface and its structure.

Figure 3 Classification of geological area

Volume 9, Issue 3, March 2020 Page 70


International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org
Volume 9, Issue 3, March 2020 ISSN 2319 - 4847

5.4 HYDROLOGY :
The dynamic detecting abilities of Radar imaging helps in hydrological contemplates. It incorporates mapping seepage
bowl, flood mapping and demonstrating of watershed and water system. It likewise gives us to evaluate soil dampness
content, snow thickness, and equivalency of snow-water.

Figure 4 Classification of hydrological area

5.5 LAND COVER AND LAND USE:


Remote detecting methods permit the mapping of land use and land in front of the earth's surface. Earlier information
ashore uses and land spread aides in overseeing common assets, ensuring natural life, observing farming and urban
exercises.

Figure 5 Classification of land area

5.6 MAPPING :
Radar information is utilized for mapping which is essential data for all remote detecting applications. It, for the most
part, incorporates Digital Elevation Model (DEM's) which gives the slant data of earth's surface and topographic
mapping or topical mapping.

Figure 6 Classification of topographic map

Volume 9, Issue 3, March 2020 Page 71


International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org
Volume 9, Issue 3, March 2020 ISSN 2319 - 4847

5.7 OCEANS AND COASTAL MAPPING:


The dynamic changes in the sea and waterfront locale can be checked and mapped utilizing remote detecting systems
that for the most part incorporate storm anticipating and sea design acknowledgment

Figure 7 Classification of coastal area

6. CONCLUSION
In this paper, we are giving input in the form of images, and using a convolution neural network algorithm for image
classification. Satellite images are the potential data source for monitoring, mapping the land area, agricultural area,
ocean, and coastal mapping, etc., Initially, Real-time data collected from Deepsat SAT-4. Then, Data must be kept in
an organized format. Finally, after processing of data and training is the very next task is obviously testing. The next
step is algorithms that are applied to data and results are noted and observed.TheDeep CNN applied to improve
accuracy at each stage.

REFERENCES
[1] Remi Ratajczak, et.al “Automatic Land Cover Reconstruction from Historical Aerial Images:An Evaluation of
features Extraction and Classification Algorithms” (2019).
[2] Gang Zheng JianguoLiu, et.al “Automatically Locate Tropical Cyclone Centers Using Top Cloud Motion Data
Derived From Geostationary Satellite Images”(2019).
[3] SinaGhassemi, et.al “Learning and Adapting Robust Features for Satellite Image Segmentation on Heterogeneous
Data Sets”(2019).
[4] Bo Du, et.al “Object Tracking in Satellite Videos Based on a Multiframe Optical Flow Tracker”(2019).
[5] Kia Zhang, et.al “Convolution Structure Sparse Coding for Fusion of Panchromatic and Multispectral
Images”(2019).
[6] M. P. Vaishnnaveet.al "A Study on Deep Learning Models for Satellite Imagery" IJAER Volume 14, Number 4
(2019) pp. 881-887
[7] F.Hu, et.al Zhang "Transferring deep convolutional neural networks for the scene classification of high-resolution
remote sensing imagery". Remote Sensing :7(11),14680–14707.
[8] C. Tao, H. et.al "Unsupervised spectral–spatial feature learning with stacked sparse autoencoder for hyperspectral
imagery classification". IEEE Geo science and Remote Sensing Letters :12 (12), 2438–2442.
[9] X. Yao, et.al "Semantic annotation of high resolution satellite images via weakly supervised learning".IEEE
Transactions on Geo science and Remote Sensing :54 (6) ,3660–3671.
[10] Y. Yang et.al "Bag-of-visual-words and spatial extensions for land-use classification".in Proc. ACM
SIGSPATIAL Int. Conf. Adv. Geogr.Inform . Syst:270-279.

Volume 9, Issue 3, March 2020 Page 72


International Journal of Application or Innovation in Engineering & Management (IJAIEM)
Web Site: www.ijaiem.org Email: editor@ijaiem.org
Volume 9, Issue 3, March 2020 ISSN 2319 - 4847

[11] Y. Wang, L. et.al "Mathiopoulos A ThreeLayered Graph-Based Learning Approach for Remote Sensing Image
Retrieval". IEEE Trans. Geosci. Remote Sens.:54(10),6020-6034.
[12] L. Gueguen et.al "Classifying compound structures in satellite images: A compressed representation for fast
queries". IEEE Trans. Geosci. Remote Sens:53( 4),1803-1818.
[13] G. Sheng, W. et.al "High resolution satellite scene classification using a sparse coding based multiple feature
combination".int.J.Remote Sens:33( 8), 2395-2412.
[14] B. Zhao, et.al "The Fisher Kernel coding framework for high spatial resolution scene classification".Remote
Sensing:8(2), 157.
[15] B. Zhao, et.al "Dirichlet-derived multiple topic scene classification model for high spatial resolution remote
sensing imagery". IEEE Trans. Geosci. Remote Sens:54(4),2108-2123.
[16] Thenkalvi, B., and S. Murugavalli. "Image retrieval using certain block based difference of inverse probability and
certain block based variation of local correlation coefficients integrated with wavelet moments." Journal of
Computer Science 10.8 (2014): 1497.
[17] Suganya, T. S., and Subramaniam Murugavalli. "Feature selection for an automated ancient Tamil script
classification system using machine learning techniques." 2017 International Conference on Algorithms,
Methodology, Models and Applications in Emerging Technologies (ICAMMAET). IEEE, 2017.

AUTHOR

Dilisha G student of Panimalar Engineering College,Department of Computer Science and Engineering

Volume 9, Issue 3, March 2020 Page 73

Вам также может понравиться