Plant Classification Using CNN Model

Anusha AR*, Arpitha MC, Ashik Hatter S, Navami BG, Rashika N and Pradeep N

Department of Computer Science and Engineering, BIET College, Davangere, Karnataka, India

*Corresponding Author:
Anusha AR
Department of CSE,
BIET College,
Davangere,
Karnataka,
India,
E-mail:
anusharavikumar666@gmail.

Received Date: July 28, 2021; Accepted Date: August 11, 2021; Published Date: August 18, 2021

Citation: Anusha AR, Arpitha MC, Hatter AS, Navami BG, Rashika N, et al. (2021) Plant Classification Using CNN Model. Glob J Res Rev Vofl.8 No.6: 85

Visit for more related articles at Global Journal of Research and Review

Abstract

The agricultural sector has recognized that, for crop management to thrive, acquiring relevant information on plants is needed. Weeds have a devastating impact on crop production and yield in general. Current practice uses uniform application of herbicides leading to high costs and degradation of the environment and the field productivity. The identification and classification of weeds are of major technical and economic importance in the agricultural industry. To automate these activities, like in shape, color and texture, a weed control system is feasible. A new Deep learning method proposed in this project identifies the type of crop using CNN. In this project different data augmentation techniques have been used for improving the classification accuracies which has been discussed to increase the performance which will help in improving the validation and training accuracies and characterization of exactness of the CNN model and accomplished various results.

Keywords

CNN; Classification; Deep learning; Agricultural sector; Environment

Introduction

Machine learning

The name Machine Learning coined in 1959 by Arthur Samuel. and Tom M. Mitchell provided a widely quoted, more formal definition of the algorithms studied in the machine learning field: "A computer program is said to learn from experience E with respect to some class of tasks T and performance measure P if its performance at tasks in P, as measured by P, improves with experience E."

Machine Learning (ML) is the scientific study of algorithms and statistical models that computer systems use to perform a specific task without using explicit instructions, relying on patterns and inference instead. It is seen as a subset of artificial intelligence. Machine learning algorithms build a mathematical model based on sample data, known as "training data", in order to make predictions or decisions without being explicitly programmed to perform the task. Machine learning algorithms are used in a wide variety of applications, such as email filtering and computer vision, where it is difficult or infeasible to develop a conventional algorithm for effectively performing the task [1].

Machine learning is closely related to computational statistics, which focuses on making predictions using computers. The study of mathematical optimization delivers methods, theory and application domains to the field of machine learning. Data mining is a field of study within machine learning, and focuses on exploratory data analysis through unsupervised learning. In its application across business problems, machine learning is also referred to as predictive analytics.

Deep learning

Deep learning is a machine learning technique that is inspired by the way human brain filters information, it is basically learning from examples. It helps a computer model to filter the input data through layers to predict and classify information. Since deep learning processes information in a similar manner as a human brain does, it is mostly used in applications that people generally do. It is the key technology behind driver-less cars, that enables them to recognize a stop sign and to distinguish between a pedestrian and lamp post. Most of the deep learning methods use neural network architectures, so they are often referred to as deep neural networks [2].

Deep learning is a branch of machine learning which is completely based on artificial neural networks, as neural network is going to mimic the human brain so deep learning is also a kind of mimic of human brain. In deep learning, we don’t need to explicitly program everything. The concept of deep learning is not new. It has been around for a couple of years now. It’s on hype nowadays because earlier we did not have that much processing power and a lot of data. As in the last 20 years, the processing power increases exponentially, deep learning and machine learning came in the picture. A formal definition of deep learning is-neurons

In human brain approximately 100 billion neurons all together this is a picture of an individual neuron and each neuron is connected through thousands of their neighbours. The question here is how we recreate these neurons in a computer. So, we create an artificial structure called an artificial neural network where we have nodes or neurons. We have some neurons for input value and some for output value and in between, there may be lots of neurons interconnected in the hidden layer [3].

Neural networks

Exploratory Data Analysis (EDA) is used by data scientists Deep Learning (DL), on the other hand, is just a type of Machine Learning, inspired by the structure of a human brain. Deep learning algorithms attempt to draw similar conclusions as humans would by continually analyzing data with a given logical structure. To achieve this, deep learning uses a multi-layered structure of algorithms called neural networks [4].

The design of the neural network is based on the structure of the human brain. Just as we use our brains to identify patterns and classify different types of information, neural networks can be taught to perform the same tasks on data.

The individual layers of neural networks can also be thought of as a sort of filter that works from gross to subtle, increasing the likelihood of detecting and outputting a correct result.

The human brain works similarly. Whenever we receive new information, the brain tries to compare it with known objects. The same concept is also used by deep neural networks.

Neural networks enable us to perform many tasks, such as clustering, classification or regression. With neural networks, we can group or sort unlabelled data according to similarities among the samples in this data. Or in the case of classification, we can train the network on a labled dataset in order to classify the samples in this dataset into different categories in Figures 1-4 [5].

global-research-diagram

Figure 1: Flow diagram.

global-research-summary

Figure 2: Model summary.

global-research-classification

Figure 3: Plant classification using CNN model.

global-research-plant

Figure 4: Result of plant classification.

Materials and Methods

Proposed work

• By using this model, we can precisely classify plants.

• A convolutional neural network model is built from scratch to extract features from a given plant image and classify it to determine if a image is plant or a weed.

• A web is built where the user can upload the plant image and the result is shown on the UI.

Objectives

• To apply preprocessing techniques on image dataset.

• To extract the most prominent features and apply CNN.

• To build a model that can detect the type of the Plant.

• The main objective of this project is to help the farmers to classify the plants and weeds more accurately using a deep learning model.

Dataset

We created our own dataset by taking image samples from the internet (Table 1).

Category Train Test
Black-grass 111 images 24 images
Common wheat 110 images 26 images
Small flowered cranesbill 115 images 28 images

Table 1: Own dataset by taking image samples from the internet.

Algorithm: CNN

Convolutional Neural Networks (CNN) are one of the most popular models used today. This neural network computational model uses a variation of multilayer perceptrons and contains one or more convolutional layers that can be either entirely connected or pooled. These convolutional layers create feature maps that record a region of image which is ultimately broken into rectangles and sent out for nonlinear processing.

Advantages: Very High accuracy in image recognition problems.

• Automatically detects the important features without any human supervision.

• Weight sharing.

Disadvantages: CNN do not encode the position and orientation of object.

• Lack of ability to be spatially invariant to the input data.

• Lots of training data is required.

In our project, as per the requirement, we have chosen the following Classification model to obtain appropriate and accurate results: Convolution neural network

Different types of Neural Networks are used for different purposes, for example for predicting the sequence of words we use Recurrent Neural Networks more precisely an LSTM, similarly for image classification we use Convolution Neural Network (CNN) [6].

Layers used to build convents: A convents is a sequence of layers, and every layer transforms one volume to another through differentiable functions.

Types of layers

1. Input layer: This layer holds the raw input of the image.

2. Convolution layer: This layer computes the output volume by computing dot product between all filters and image patches.

3. Activation function layer: This layer will apply element wise activation function to the output of the convolution layer. Some common activation functions are RELU, Sigmoid, Tanh, Leaky RELU, etc. We have used RELU as well as softmax for our model as this is a multi-classification problem.

4. Pool layer: This layer is periodically inserted in the convNets and its main function is to reduce the size of volume which makes the computation fast, reduces memory and also prevents from overfitting. Two common types of pooling layers are max pooling and average pooling.

5. Dense layer: It is the regular deeply connected neural network layer

Accuracy is defined as the ratio of the number of samples correctly classified by the classifier to the total number of samples for a given test data set.

We have used Non-Linearity (ReLU) activation function. ReLU stands for Rectified Linear Unit for a non-linear operation.

The output is ƒ(x)=max (0, x).

ReLU‟s purpose is to introduce non-linearity in our ConvNet. Since, the real world data would want our ConvNet to learn would be non-negative linear values. We have used the softmax function as this is a multi-class classification problem.

Results and Discussion

Data collection

In this section, the steps that are performed to get the desired output are explained with the help of flow diagram (Figure 1). The dataset we have used for this project is we created our own dataset by taking image samples from the internet. Our dataset consist of 414 images which are divided into train and test dataset.

Pre-processing

The dataset is pre-processed to eliminate disturbance and cleaning of data is done through this method. Later the pre- processed data is trained.

Defining model architecture

For image classification we use Convolution neural network. This is a very crucial step in our deep learning model building process. We have to define how our model will look and that requires

Importing the libraries, initializing the model, Adding CNN (Convolution Neural Network) Layers, Adding Dense layers

Configure the learning process

With both the training data defined and model defined, it's time configure the learning process. This is accomplished with a call to the compile() method of the Sequential model class.

Train and test data

For training a model we initially split the model into 2 sections which are ‘Training data’ and ‘Testing data’. The classifier is trained using ‘training data set’, and then tests the performance of classifier on unseen ‘test data set’.

Train and save the model

At this point we have training data and a fully configured neural network to train with said data. All that is left is to pass the data to the model for the training process to commence, a process which is completed by iterating on the training data. Training begins by calling the fit() method. Your model is to be saved for the future purpose [7,8].

Model

• tf.keras.layers.Conv2D (): The convolution layer which improves image recognition by isolate images features (Figure 2).

• tf.keras.layers.MaxPooling2D (): a layer to reduce the information in an image while maintaining features.

• tf.keras.layers.Flatten (): flatten the result into 1-dimensional array.

• tf.keras.layers.Dense (): add densely connected layer.

• Loss: pneumonia detection is using sigmoid activation in the final step, which resulted in either 0 or 1 (normal or pneumonia). Therefore, binary cross entropy is the most suitable loss function.

• Metrics: accuracy is the measurement metric to obtain the prediction accuracy rate on every epoch

Conclusion

In the given study, it has been demonstrated that how one can classify the plants and weeds easily from a small dataset of images. This model was basically built from the scratch that helps in separating it from all other existing methods like transfer learning etc. When the neural networks were trained using datasets containing a three species, it performed well to recognise each variety of plants. As the result, the accuracy of the test dataset reached 98.07%, loss is 93.30%, val_loss is 45.13% and val_accuracy is 83.33% which indicates a decent model. Therefore, possible solutions for considering these problems are great area for the research domain.

References

Select your language of interest to view the total content in your interested language

Viewing options

Flyer image
journal indexing image

Share This Article