fbpx

Spectral Grouping For Image Classification


Call for Papers Engineering Journal, May 2019

Download Full-Text PDF Cite this Publication

Text Only Version

Spectral Grouping For Image Classification

1 B. Naga Jyothi, 2 K.S.R. Radhika,

1 Assoc. Prof., Dept. of ECE, DMS SVHCE, Machilipatnam, A.P., India, jyothi_9802@yahoo.com

2 Assoc. Prof., Dept. of CSE, DMS SVHCE, Machilipatnam, A.P., India, kammilisrr@gmail.com

Abstract: This paper presents a spectral based approach for image classification. The method divides the image into several groups by determining the closeness of each image pixel to each of the samples selected from the image using mahalnobis distance as a measure of similarity. Each pixel is then assigned to its closest sample. The result is an image composed of sample IDs.

Keywords: Image Classification, Pixel Labelling, Mahalanobis Distance

  1. Introduction

    Based on the image primitive used viz., pixel based or object based, image classification methods are broadly of two main categories. Pixel based method as described in this paper classify individual pixels using only the spectral patterns. Object based methods use the spectral, spatial and contextual information of the objects to perform image Classification.

    The input image is Landsat 7 pan image of Pyongyang, North Korea acquired on September 17, 2002 [Ref.1, www.nasa.gov].

    Fig:1 Input Image

    This is a natural-color image using ETM+ bands 3, 2, 1. In this image, Pyongyang appears grey in color and is surrounded by vegetation (in green). The Taedong River, shown in dark blue, travels through the city.

  2. Methodology

    Samples are selected based on the expected. Number of classes in the image. Spectral grouping is done by determining the closeness of each image pixel to each of the samples selected from the image.The pixel is assigned to its closest sample based on Mahalanobis distance measure of closeness.

    The Mahalanobis distance D as defined below, is used a measure of closeness or similarity.

    D= (x-µ) (x-µ)T ——— (

  3. Implementation

    The programming has been done using

    MATLAB 7.0.1 [ Ref.2 Lonesome M.Malambo 2009]. The sequence of steps involved in the developed code are

    • The image data is loaded using MATLABs imread function[Fig:1]

    • RGB is converted to double format to facilitate floating point calculations.

    • The image is smoothed using a Gaussian filter to reduce noise. fspecial command together with imfilter command are used to achieve this.

    • Individual bands are separated.

    • Samples are selected from different classes Viz., vegetation, water, settlement and barren regions[Fig:2]

      MATLABs built in function roipoly is used to select regions of interest.

      roipoly returns a binary image that can be used as a mask for masked filtering.

    • Statistics of the samples (mean & covariance) are calculated.

      MATLABs mean and cov functions are used for calculating the mean and covariance respectively.

    • The samples are Labelled (ID no. is given)

    • Mahalanobis distance is then calculated between each pixel and each instance of sample set. pixels close to a sample are assigned the same sample ID.

    • The result is an image composed of sample IDs

    • Finally the classified image is observed using MATLABs reshape function.

  4. Results

    Fig:2 selected samples using roipoly function

    1 )

    In (1), x is the pixel spectral vector, is the mean spectral vector of a sample in a multiband image, is the covariance matrix of the sample, T denotes the transpose of the matrix.

    Fig:3 selected samples superimposed on the input image

    Sample 1: Vegetation , Sample 2: water Sample 3: Settlement, Sample 4: Barren

    Fig: 4 output Classified image

    Vegetation: dark blue Water: light blue Settlement: yellow Barren: Red

    Mean and Co-Variance of vegetation sample:

    0.4355

    0.6661

    0.3177 V1N2-181-190,2012

    [7].KSR Radhika, B.Naga Jyothi, et.al.Accuracy Assessment

    0.0087

    0.0083

    0.0086

    of per pixel Based Classification Conference Proceedings, NC-

    0.0083

    0.0140

    0.0093

    Velasiem-2k12: Pg.73-77

    0.0086

    0.0093

    0.0128

    nd Co-Va

    riance of

    water sa

    mple: [8] MathWorks.Image Processing Toolbox 6.2. Mathworks.com

    —————————————–

    0.4376 0.4813 0.4533

    0.0292 0.0256 0.0239

    0.0256 0.0228 0.0213

    0.0239 0.0213 0.0205

    Mean and Co-Variance of barren sample:

    ————————————- 0.6508 0.6743 0.6184

    0.0281 0.0273 0.0296

    0.0273 0.0283 0.0293

    0.0296 0.0293 0.0326

  5. Conclusion

Good results have been obtained as can be seen by comparing the input and output classified images. The classification result is influenced by the samples and also the number of samples. However pixel based methods suffer from mixed pixel problem. This can be overcomed by object based methods which uses spectral, spatial and contextual information of the objects to perform image Classification.

References

[1]. www.nasa.gov + NASA Home Multimedia. http://landsat.gsfc.nasa.gov/images/archive/c0018.html

[2]. Lonesome M.Malambo A region based Approach to Image Classification ,Applied Geoinformatics for Society and Environment 2009-Stuttgart University of Applied Sciences

[3]. Schowengerdt, Robert A. (1997). Remote Sensing: Models and Methods for Image Processing (2nd ed.). New York: Academic Press.

[4]. D. LU and Q. WENG A survey of image classification methods and techniques for improvingclassification performance

International Journal of Remote Sensing Vol. 28, No. 5, 823 870, 10 March 2007

[5]. John R.Jensen Introductory Digital Image Processing, A remote Sensing perspective, Third Edition, PPH

[6].B.Naga Jyothi, Dr. G. R. Babu, Dr.I.V.Murali Krishna, Thematic classification of multispectral imagery, International Journal of Electronics and computer Science Engineering ,

Mean a

———————————— 0.0474 0.1330 0.1545

0.0106 0.0100 0.0095

0.0100 0.0095 0.0091

0.0095 0.0091 0.0088

Mean and Co-Variance of settlement sample:

—————————————–

Acknowledgements

"Landsat imagery courtesy of NASA Goddard Space Flight Center and U.S. Geological Survey" or "USGS/NASA Landsat" for the images.

Leave a Reply

Your email address will not be published. Required fields are marked *