Analysis of Plant Leaves Disease Classification Using Feed Forward and Cascade Forward Neural Network

DOI : 10.17577/IJERTCONV3IS16094

Download Full-Text PDF Cite this Publication

Text Only Version

Analysis of Plant Leaves Disease Classification Using Feed Forward and Cascade Forward Neural Network

K. Muthukannan1

Research Scholar,

Department of Electronics and Communication Engineering, Einstein College of Engineering,

Tirunelveli-627 012, Tamilnadu, India.

P. Latha2

Associate Professor,

Department of Computer Science and Engineering, Government College of Engineering, Tirunelveli-627 007, Tamilnadu, India.

Abstract– This paper describes the analysis of plant leaves dis- ease classification using feed forward neural network (FFNN) and cascade forward neural network (CFNN). Features play a vital role for any classification which is one of the deciding fac- tors for the performance of classification accuracy. Ini- tially, the features for classification is extracted from the seg- mented image after the processing of segmentation using Parti- cle swarm optimization (PSO).This paper mainly focus on ex- traction of features such as color, shape and texture. Then, the extracted features are trained by the neural networks such as FFNN and CFNN and also the performance is analyzed for different disease affected plant leaves using confusion ma- trix. Experimental results demonstrate the performance of the proposed approach producing comparable classification accuracy of plant leaves with disease.

Keywords- Features extraction, color image segmentation, particle swarm optimization, feed forward neural network, cas- cade forward neural network, confusion matrix.

  1. INTRODUCTION

    An image may be defined as a two dimensional func- tion f(x, y), where x and y are spatial (plane) coordinates, and the amplitude of f at any pair of co-ordinates (x, y) is called the intensity or gray level of the image at that point [10], [15]. The region of interest in the image can be degraded by the impact of imperfect instrument, the problem with data acquisition process and interfering natural phenomena. Therefore, the original image may not be suitable for analysis. Thus image segmentation technique is often neces- sary and should be taken as significant step during image is processed and analyzed. Repeatable experiments with published benchmarks are required for this research field to progress. Choosing an appropriate model for seg- mentation is a difficult task to achieve better segmentation with reduced computational time. The problem is reformed with minimized computational time and high quality of the results.

    The classification of unnatural plant leaves is a crucial process in botany and other industries. Moreover, the morphological features of leaves are used for plant classifi- cation or in the early diagnosis of certain plant diseases [3]. This paper presents the design and implementation of an artificial vision system capable of extracting HFE fea- tures from unnatural plant leaves. Initially, leaves taken from plants leaves are collected from nearby surroundings

    and used as samples for testing the proposed system. Later, additional samples originating from diverse environments are used for classification [8]. The proposed system consists of:

    a) an artificial vision system (camera) b) a combination of image processing algorithms such as filtering, segmentation using PSO and extraction of features are implemented in Matlab and c) a feed-forward neural network (FFNN) and cascade forward neural network (CFNN) based classifier implemented in Matlab [4],[5]. The image processing part is responsible for image capture and image pre-processing in order to obtain normalized features [6], [7] and for deter- mining some critical geometrical characteristics. The HFE feature has been extensively used in this proposed work.

    Artificial neural networks are biologically stimulated classification algorithms that consist of an input layer of nodes, one or more hidden layers and an output layer [20]. Each node in a layer has one corresponding node in the next layer, thus creating the stacking effect [1]. Artificial neu- ral networks are the very adaptable tools and have been widely used to tackle many issues. Feed-forward neural net- works (FNN) are one of the admired structures among artifi- cial neural networks [14]. These well-organized networks are widely used to solve complex problems by modeling com- plex input-output relationships. Neural networks are well known as powerful tools in the area of pattern classification [2]. In principle, multi-layer feed-forward networks with just a single hidden layer is universal approximators for arbitrary finite-input environment measures. However, It does not imply that a neural network can easily be trained the under- lying functional mapping between the input data and the de- sired output. In fact, the most important drawbacks of neural networks are problems associated with local minima and the slow convergence of the learning process. To tackle these problems, neural networks are used to reduce learning com- plexity [12].

    The paper is organized as follows, In Section II shows the problem statement and data description .Materials and methods are described in Section III. In Section IV, a brief description of the different classifiers and its perfor- mance measures are discussed and the Section V contains the summary of conclusion.

  2. PROBLEM STATEMENT AND DATA DESCRIPTION

    This aim of the research work is to evaluate the perfor- mance of unnatural plants data set classification using FFNN and CFNN, which is described as follows:

    Proposed unnatural plants data set: The new data base is named as unnatural plant leaves data set. It has bitter guard, beans, chilly, cotton, pigeon pea and tomato leaves images and it is considered for this research. It consists of 470 image databases which are taken from the field by a digital camera under natural light condition with certain specification. This data set contains 6 classes, where each class refers to a type of unnatural plant leaves of six different plants. The data base contains the following features [18].

    1. Texture features: It contains four attributes such as contrast, correlation, energy and homogeneity .

    2. Shape features: It contains five attributes such as area, perimeter, major axis length, minor axis length and eccen- tricity.

    3. Color features: It contains twelve attributes such as mean_ R, mean_ G, Mean_ B, Variance_ R, Variance_ G, Variance_ B, Standard deviation_ R, Standard deviation_ G, Standard deviation_ B, median_ R, median_ G, median_ B.

    4. Classes: Unnatural plant leaves of bitter guard, beans, chilly, cotton, pigeon pea and tomato.

      The goal of this research is to demonstrate the process of building a neural network based classifier that solves the classification problem and compare performance of the clas- sifiers such as feed forward neural network (FFNN) and cas- cade forward neural network (CFNN) using confusion matrix and other performance metrics such as sensitivity, specificity, precision ,recall and F_measure.

      The images used in this work were taken with the following image acquisition system. Sony Cyber-Shot Dsc-W830 Point & Shoot digital camera with 20.1 Mega Pixels of resolution, placed vertically at a distance of 20 cm from the samples. The angle between the axis of the lens and the sources of illumination is approximately 45degree. The images were taken at maximum resolution (4608 x 3456 pixels). Fig.1 shows the functional block diagram of the proposed unnatu- ral plant leaves data set (features) extraction process. Here more numbers of disease infected plant leaves data bases are collected by means of image acquisition [6]. It is then ap- plied for image segmentation using particle swarm optimiza- tion (PSO) after the process of image preprocessing. The features such as texture features from the gray level co-occurance matrix, shape features from the binary image of the original image and color features from the R, G and B individual plane of the original image is extracted [1]. After extracting the features, the best features is to be find for bet- ter classification and the performance measures are analyzes with the help of confusion matrix as well as some metrics.

      1. Method for Feature Extraction

        Gray level co-occurance matrix (GLCM) is a method for texture feature calculation [9]. When the GLCM is generated, there are a total of 14 textures features that could be com- puted from the GLCM, such as contrast, variance, sum aver- age etc [19].The four common textures features discussed here are contrast, correlation, energy, and homogeneity. Contrast is used to measure the local variations.

  3. MATERIALS AND METHODS

    Analysis the performance

    Training/ Testing inputs for any Classifiers

    Different disease affected leaf samples

    Input Image

    Preprocessing (Filtering by Gaussian)

    Segmentation (Particle Swarm Optimization)

    Feature Extraction

    Classifiers (FFNN&CFNN)

    Fig. 1 Proposed Description Model- Plant Leaves Disease Classification

    Correlation is used to measure the probability of occur- rence for a pair of specific pixels. Energy is also known as uniformity of ASM (angular second moment) which is the sum of squared elements from GLCM, and homogeneity is to measure the distribution of elements in the GLCM with re- spect to the diagonal. A gray-level co-occurrence matrix (GLCM) is essentially a two-dimensional histogram in which the (i, j) th element is the frequency of event i co-occurs with event j. A co-occurrence matrix is specified by the relative frequencies P (i, j, d, ø) in which two pixels, sepa- rated by distance d occur in a direction specified by the angle

    ø, one with gray level i and the other with gray level j. A co-occurrence matrix is therefore called as a function of dis- tance r, angle ø and grayscales i and j [13].

    1. Particle Swarm Optimization (PSO)

      Particle swarm optimization (PSO) is a population based stochastic optimization technique developed by Dr. Eberhart and Dr. Kennedy in (1995), inspired by social be- havior of bird flocking or fish schooling.PSO is initialized with a group of random particles (solutions) and then

      searches for optima by updating generations. In every itera- tion, each particle is updated by following two "best" values. The first one is the best solution (fitness) it has achieved so far. (The fitness value is also stored.) This value is called pbest. Another "best" value that is tracked by the particle swarm optimizer is the best value, obtained so far by any particle in the population. This best value is a global best and called gbest. When a particle takes part of the population as its topological neighbors, the best value is a local best and is called lbest.

    2. Multilayer Feed Forward Back Propagation Neural Net- work

      Feed forward networks often have one or more hidden layers of sigmoid neurons followed by an output layer of linear neurons. Multiple layers of neurons with nonlinear transfer functions allow the network to learn nonlinear and linear relationships between input and output vectors [1]. The linear output layers of the network produce values outside the range -1 to +1. Network architecture is determined by the number of hidden layers and by the number of neurons in each hidden layer. The network is trained by the back propa- gation learning rule.

      To build a neural network, it is sufficient to combine the neural layers. Each layer has its own matrix weight Wk, where k designates the index of the layer. Thus, the vectors bk, nk and ak are associated to the layer k. To specify the neu- ral network structure, the number of layers and the number of neurons in each layer must be chosen. The learning step is a dynamic and iterative process which consists in modifying the parameters of the network after receiving the inputs from its environment. The learning type is determined by the way the change of parameters occurs [14]. Most of the neural

      In the presence of sample-data combinations en- tries/outputs {(, )}, = 1 , where designates an entry-vector and desired output-vectors, we can for- ward propagate at each instant t, an entry-vector

      () through the neural network in order to get an out- put-vector ().

      Given here are (), the error produced by the network calculation for an entry, and the corresponding desired out- put ():

      () = () () (3)

      The performance function F that permits minimizing the root mean square error is defined by the following expres- sion:

      () = [(). ()] (4)

      Where E [ ] and x are respectively the mean and the vector grouping the set of the weights and the bias of the neural network. F is approximated on a layer with the in- stantaneous error:

      () = (). () (5)

      The method of the steepest falling gradient is used to optimize X with the help of the following equations:

      (6)

      where designates the learning rate of the neural net- work. To calculate the partial derivatives of the rule of composition functions is used

      architectures encountered, the learning results in the synaptic

      =

      .

      and

      modification of the weights connecting one neuron to the

      other. If

      () designates the weight connecting the neu-

      =

      .

      (7)

      ron i to its entry j and at time t, a change () of the weights can simply be expressed by the Equation (1):

      () = ( + 1) () (1) Where ( + 1)represents the new entry value of the

      The activation levels of the layer k depend directly

      on the weights and biases on this layer, and can be expressed by the following relation:

      1

      = 1 ,

      weight .

      A set of well-defined rules that allow realizing a weight adaptation process is designated by a learning algorithm of the neural network.

      =1

      = 1,2, . (8)

      Now, for the first terms of the Equation (9), we define the sensitivities of in relation with the changes of the activation level of the neuron i belonging to the layer k

    3. The Feed Forward Back Propagation Method Used by a Multilayer Neural Network

      by the equation:

      =

      (9)

      The expressions of () and () become:

      Let us consider the multilayer neural network (com- posed of M layers)[21]. The equation describing the outputs of the layer k is:

      = (1 ), = 1,2, . (2)

      () = . (). (1)()

      and

      () = . () (10) Equation (10) is responsible for the modification of the

      weight connections and biases and the different iteration learning steps of the neural network presented in Section 3.3.1.

    4. The Cascade Forward Back Propagation Method Used by a Multilayer Neural Network

      By considering different analyzes, we will give an ab- stract of the methodology used in the learning process that we have implemented [17].

      1. Initialize the weights with small random values;

      2. For each combination (, ) in the learning sample:

        Propagate the entries forward through the neural network layers:

        The implementation starts with initialize the training data- base then resize the images of size 256*256. Next Gaussian special filter was used to remove the noise and to improve the iage quality. Features extracted from the test image (new features) are compared with the features available in the training set (features). Texture, Shape and color features are extracted and stored as inputs [7] [11]. We have evaluat- ed the mean squared error (mse) by using equation (15).

        (15)

        Where Ei: Error (observed valueNeural Network estimate); and N: Number of observed values.

        0 = ;

        = (1 ),

        = 1,2, . (11)

        Back propagate the sensitivities through the neural net- work layers:

        = 2()( );

        = ()(+1)+1,

        = 1, 1 (12)

        Modify the weights and biases:

        = (1), = 1, (13)

        = , = 1, (14)

      3. If the stopping criteria are reached then stop; if not reached, they permute the presentation order of the combina- tion built from the learning database, and begin again at Step 2.

    5. Cross-Validation

    Cross-validation procedure is applied to provide better generalization of neural network classifiers. To perform the cross-validation procedure, the input data is partitioned into 3 sets: 1) Training set 2) Validation set 3) Test set

    The training set is used to train the network. The valida- tion set is used to validate the network and to adjust network design parameters. The test set is used to test the generaliza- tion performance of the selected design of neural net- work. The partitioning of input data is performed ran- domly with a certain ratio of input entities to be stored as training set, validation set and test set (0.8, 0.1 and 0.1 re- spectively).

  4. RESULTS AND DISCUSSIONS

    The disease affected leaves dataset are collected from south zone of Tamil Nadu at Pavoorchatram in the district of Tirunelveli in the month of July to September 2014.The in- fected plant leaves are collected and kept in a plastic cover for image acquisition. Here, the data base size is 470 im- ages which are collected from the field.

    In this research, the dataset is captured and collected by camera (Sony) in the above said areas. Six types of disease affected plant leaves such as Bittergaurd, Beans, Chilly, Cotton, Pigeon and Tomato were used for analysis [16],[21].

    1. Experimental Results and Analysis for Feed Forward Back Propagation Neural Network

      The analysis of feed forward back propagation neural net- work (FFNN) results from confusion matrix is based on the features of texture, shape and color features of different classes of unnatural plant leaves and results are shown in Fig,2.

    2. Results and Analysis of FFNN Using Confusion Matrix for Different Features of Class=6

      The classification result for the confusion matrix obtained under positive versus negative classification model is shown in Fig. 2. The overall classification accuracy of FFNN is 57.2%, 50.9% and 59.1% for texture, shape and color and its corresponding error rate is 42.8% ,49.1%, 40.9% for Class=6. For example in Fig.2 (a), the true positive value for unnatural bitterguard, beans, chilly, cotton, pigeon pea and tomato leaves are 29.1%, 3.3%, 44.4%, 27.3%, 93.4% and 84.6%

      respectively and its corresponding false rate of the unnatural leaves are 70.9%, 96.8%, 55.6%, 72.7%, 6.6% and 15.4%.Then, the error rate is calculated by using mean squared error (mse) and the best validation performance of the mean squared error (mse) is

      1.5765 at one epoch and is shown in Fig. 2(b).

      Fig. 2(a) Results of the confusion matrix of performance analysis of FFNN a) texture, class=6,(b) Results of the MSE (training, validation and testing)

      Similarly the overall classification accuracy, error rate of classification, true positive and false rate for individual in- fected plant leaves are calculated from the confusion matri- ces. From these confusion matrix analyzes, the performance of the FFNN is very low for texture features and clearly says the better features for better classification.

    3. Experimental Results and Analysis for Cascade Forward Back Propagation Neural Network

      The analysis of cascade forward back propagation neural network (CFNN) results from confusion matrix is based on the features of texture, shape and color features for different classes of unnatural plant leaves that results are shown in section 4.2.1 [17].

    4. Results and Analysis of CFNN Using Confusion Matrix for Different Features of Class=6

      The classification result for the confusion matrix obtained under positive versus negative classification model is shown in Fig.3.

      The overall classification accuracy of CFNN is 60%, 50.2% and 62.8% for texture, shape and color and its corresponding error rate is 40% ,49.8%, 37.2% for Class=6. For example in Fig.3 (a), the true positive value for unnatural bitterguard, beans, chilly, cotton, pigeon pea and tomato leaves are 29.1%, 4.8%, 46.7%, 40.9%, 94.6% and 88.5% respectively and its corresponding false rate of the unnatural leaves are 70.9%, 95.2%, 53.3%, 59.1%, 5.4% and 11.5%.Then, the

      error rate is calculated by using mean squared error (mse) and the best validation performance of the mean squared error (mse) is 1.5765 at one epoch which is shown in Fig.3(b). Similarly, the overall classification accuracy, error rate of classification, true positive and false rate for individu- al infected plant leaves are calculated from confusion ma- trices. From these confusion matrix analyzes, CFNN gives

      the optimal results compared to FFNN based classification. Table 2, demonstrates the results of evaluation of the per- formance measures for classifier FFNN and CFNN of disease affected leaves for different features, Class=6.

      Fig. 3(a) Results of the confusion matrix of performance analysis of CFNN a) texture, class=6, (b) Results of the MSE (training, validation and testing)

    5. Comparison of the Result Performance between FFNN and CFNN

    The performance comparison of feed forward back prop- agation neural network (FFNN) and cascade forward back propagation neural Network (CFNN) is based on the features of texture, shape and color features for class=6 of unnatural plant leaves that results are shown table.1

    Table 1 Analysis of features (texture, shape and color), class=6

    Performance metrics

    Techniques

    Bitterguard (Class1)

    Beans (Class2)

    Chilly (Class3)

    Cotton (Class4)

    Pigeon pea (Class5)

    Tomato (Class6)

    Accuracy

    FFNN

    0.127

    0.175

    0.678

    0.659

    0.97

    0.923

    CFNN

    0.309

    0.397

    0.867

    0.864

    0.976

    0.942

    Sensitivity

    FFNN

    0.127273

    0.174603

    0.677778

    0.659091

    0.96988

    0.923077

    CFNN

    0.309091

    0.396825

    0.866667

    0.863636

    0.975904

    0.942308

    Specificity

    FFNN

    0.875

    0.578947

    0.717647

    0.707317

    0.755869

    0.461538

    CFNN

    1

    0.675676

    0.722222

    0.77551

    0.84375

    0.731343

    Precision

    FFNN

    0.875

    0.578947

    0.717647

    0.707317

    0.755869

    0.461538

    CFNN

    1

    0.675676

    0.722222

    0.77551

    0.84375

    0.731343

    Recall

    FFNN

    0.127273

    0.174603

    0.677778

    0.659091

    0.96988

    0.923077

    CFNN

    0.309091

    0.396825

    0.866667

    0.863636

    0.975904

    0.942308

    F_measure

    FFNN

    0.222222

    0.268293

    0.697143

    0.682353

    0.849604

    0.615385

    CFNN

    0.472222

    0.5

    0.787879

    0.817204

    0.905028

    0.823529

    Table 2 Analysis of classification accuracy for features, class=6

    No of Class

    Features

    Accuracy of Classifiers (%)

    FFNN

    CFNN

    6

    Texture

    57.2

    60

    Shape

    50.9

    50.2

    Color

    59.1

    62.8

    Fig.4 Results of the Comparison of Performance analysis of FFNN and CFNN using (Texture, shape and color) features for class=6, 3and 2

    The Fig.4 shows the comparison of the results of FFNN and CFNN. This result clearly mentioned that the classifica- tion accuracy of classifiers using color features gives optimal results comparably to other features.

  5. CONCLUSIONS

In this paper, the disease infected plant leaves classifi- cation is presented and classified by using back propagation algorithm based FFNN and CFNN. Here, the image segmen- tation was done by PSO. After segmenting the disease af- fected portion of the leaves, the texture features were ex- tracted by using gray level cooccurance matrix. Similarly, shape features extracted from the binary image of the data- base and color features were extracted from the individual R, G, and B color plane of the original image. Then the perfor

mances of the neural networks were analyzed by using texture, shape and color features separately. According to the performance measurement indicators, the CFNN is indeed slightly superior to the FFNN under certain conditions. Here the classification accuracy of CFNN using color feature is better than other features.

As the extension of this work, the hybrid features based plant leaves disease classification will be considered for our research work and also it will produce better results for clas- sification compared to the current features.

ACKNOWLEDGMENT

The authors would like to thank the reviewers for their valuable suggestions which help them for improving the quality of this paper. We would like to thank the Internation- al Rice research Institute, Tamilnadu Agricultural University, Coimbatore and also the management of Einstein College of Engineering for their continuous encouragement and support.

REFERENCES

  1. N.E.Abdullah,A.A. Rahim,Classification of Rubber Tree Leaf Diseases Using Multilayer Perceptron Neural Network, 5th Student Conference on Research and Development, DOI: 10.1109 ,2007, pp. 1 6.

  2. Adrian Ford, Alan Roberts, Colour Space Conversions, August11, 1998.

  3. H.Al-hiary, bani-ahmad, Fast and Accurate detection and classifica- tion of plant diseases, International Journal of Computer Applica- tions (0975-8887), March 2011, Volume 17-No.1.

  4. A.Akhtar, A.Khanum,"Automated Plant Disease Analysis (APDA): Performance Comparison of Machine Learning Techniques, 11th In- ternational Conference on Frontiers of Information Technology (FIT), DOI: 10.1109/FIT.2013.19,2013,pp.60 65.

  5. D. Al Bashish, M. Braik, "A framework for detection and classification of plant leaf and stem diseases,2010 International Conference on Signal and Image Processing (IC- SIP),DOI: 10.1109/ICSIP.2010.5697452 , 2010 , pp.113 118.

  6. G. Anthonys, N. Wickramarachchi, An image recognition system for crop disease identification of paddy fields in Sri Lanka, International Conference on Industrial and Information Systems (ICIIS), DOI: 10.1109/ ICIINFS. 2009.5429828 , 2009, pp. 403 407.

  7. Andrzej Materka, Michal Strzelecki, Texture Analysis Methods – A review, technical university of Lodz, Institute of Electronics, COST B11 report, Brussels 1998.

  8. H. Fu, Z. Chi, "Combined thresholding and neural network approach for vein pattern extraction from leaf images, IEE Proceeding – Vision, Image and Signal Processing, Vol.153, Issue:6 ,DOI: 10.1049/ip-vis:20060061, 2006 ,pp. 881 892.

  9. R. C. Gonzalez and R. E. Woods, Digital image processing, Pearson Education, 2002.

  10. HangZhang,P.Yanne,"Plant Species Classification Using Leaf Shape and Texture, International Conference on Industrial Control and Electronics Engineering (ICICEE), DOI: 10.1109/ ICICEE.2012.538 , 2012 , pp. 2025 2028.

  11. Hao Gao , Wenbo Xu, Multilevel Thresholding for Im- age Segmentation Through an Improved Quantum-Behaved Particle Swarm Algorithm, IEEE Transactions on Instrumentation and Measurement, Vol.59, Issue:4,DOI:10.1109/TIM.2009.2030931,2010 , pp. 934 946.

  12. Hum yan chai, Gray-Level-Co-occurance Matrix Bone fracture De- tection, WSEAS transaction on Systems, Issue 1, Vol.10, January 2011.

  13. D.Hush,Classification with neural networks: a performance analysis, IEEE International Conference on Systems Engineering, DOI:10.1109/ICSYSE.1989.48672 ,1989 , pp.277 280.

  14. R. Janani, A. Gopal, Identification of selected medici- nal plant leaves using image features and ANN , International Con- ference on Advanced Electronic Systems (ICAES), DOI: 10.1109/ ICAES. 2013. 6659400, 2013, pp. 238 242.

  15. S.B.Kutty, N.E. Abdullah, "Classification of watermelon leaf diseas- es using neural network analysis, Business Engineering and Industrial Applications Colloquium (BEIAC), IEEE , DOI: 10.1109/ BEIAC. 2013. 6560170 , 2013, pp. 459 464.

  16. S. Phadikar, J. Sil, Rice disease identification using pattern recogni- tion techniques, 11th International Conference on Computer and In- formation Technology, ICCIT 2008, DOI: 10.1109/ICCITECHN. 2008. 4803079 , 2008, pp. 420 423.

  17. Prasetiyo ,M.Khalid, A Comparative Study of Feature Extraction Methods for Wood Texture Classification, Sixth International Con- ference on Signal-Image Technology and Internet-Based Systems (SI- TIS), DOI:10.1109/,2010,pp.23 29.

  18. P. Laird and R. Saul, Automated feature extraction for supervised learning, IEEE World Congress on Computational Intelligence, Proceedings of the First IEEE Conference on Evolutionary Computa- tion, DOI: 10.1109/ICEC.1994.349977, Vol.2,1994 , pp.674 679.

  19. R.P.Lippmann,Pattern classification using neuralnetworks,Communic ations Magazine,

    IEEE Vol.27, Issue:11DOI: 10.1109/35.41401,1989 ,pp. 47 50.

  20. R. Pydipati, T. Fburks, Statistical and neural network classifier for citrus disease detection using machine vision, Transaction of the ASAE, vol.48 (5):2007-2014.

  21. S.S. Sannakki, V.S. Rajpurohit,Diagnosis and classification of rape leaf diseases using neural networks, Fourth International Con- ference on Computing, Communications and Networking Technologies (ICCCNT), DOI: 10.1109/ICCCNT.2013.6726616, 2013 ,pp.1- 5.

Leave a Reply