Corrected Main
Corrected Main
PREAMBLE
1.1 INTRODUCTION
As we know since the humans have started practicing or doing agriculture activities
“Agriculture” has become the most important activity for humans. In today’s era or world
agriculture is not only for surviving it’s also play huge part or role in economy of any
country. Agriculture plays vital role in India’s economy and in human future too. In India it
also provides large portion of employment for Indians. As a result, with passage of time the
need for production has been accumulated exponentially. thus, on manufacture in mass
amount individuals are exploitation technology in associate degree extremely wrong method.
With the improvement of the technologies day by day there is creation of hybrid
varieties day by day. In comparison with naturally created crop these hybrid varieties don’t
offer or provide essential contents. Depending more on unnatural techniques may lead to soil
acidification and crust. These types of activities all lead up to environmental pollution. These
types of activities (unnatural activities) are for avoiding or reducing losses. However, once
the farmer or producer get or grasp the correct data on the crop yield, it will help the farmer
in avoiding or reducing the loss.
Around the globe India is the second largest country in terms of population. Many
people are dependent on agriculture, but the sector lacks efficiency and technology especially
in our country. By bridging the gap between traditional agriculture and data science, effective
crop cultivation can be achieved. It is important to have a good production of crops. The crop
yield is directly influenced by the factors such as soil type, composition of soil, seed quality,
lack of technical facilities etc.
Agriculture sector act as backbone of India by providing food security and playing
major role in Indian economy. Due to drastically changes in climatic condition it is affecting
farmers due to poor yield, which also affect them economically. Due to it prediction of crop
is getting difficult for farmers. This project will help the upcoming farmers by giving the
farmer ease in predicting the crop to sow for maximum profit.
In India agriculture plays important role in economic sector and also plays the most
important role in global development. A lot more than 60% of the country's land is used for
agriculture to meet the needs of 1.3 billion people. So, adopting new technologies for
agriculture plays important role. This is going to lead our country's farmers to make a profit.
Crop prediction and fertilizer prediction in most part of part India is done on by the farmers
experience. Most farmers will prefer previous or neighboring crops or most prone to the
surrounding region only because of their land and do not have sufficient information about
the content of soil like phosphorus, potassium, nitrogen.
1
1.2 EXISTING SYSTEM
More and more researchers have begun to identify this problem in Indian agriculture
and are increasingly dedicating their time and efforts to help alleviate the issue. Different
works include the use of Regularized Greedy Forest to determine an appropriate crop
sequence at a given time stamp. Another approach proposes a model that makes use of
historical records of meteorological data as training set. Model is trained to identify weather
conditions that are deterrent for the production of apples. It then efficiently predicts the yield
of apples on the basis of monthly weather patterns. The use of several algorithms like
Artificial Neural Network, K Nearest Neighbors, and Regularized Greedy Forest is
demonstrated in [5] to select a crop based on the prediction yield rate, which, in turn, is
influenced by multiple parameters. Additional features included in the system are pesticide
prediction and online trading based on agricultural commodities.
1.2.1 DRAWBACKS
One shortcoming that we identified in all these notable published works was that the
authors of each paper concentrated on a single parameter (either weather or soil) for
predicting the suitability of crop growth. However, in our opinion, both these factors should
be taken together into consideration concomitantly for the best and most accurate prediction.
This is because, a particular soil type may be fit for supporting one type of crop, but if the
weather conditions of the region are not suitable for that crop type, then the yield will suffer.
2
1.4 PLAN OF IMPLEMENTATION
The steps involved in this system implementation are:
b) Data Preprocessing: This step includes replacing the null and 0 values for yield by -1
so that it does not effect the overall prediction. Further we had to encode the data-set so
that it could be fed into the neural network.
c) Training model and crop recommendation: After the preprocessing step we used the
data-set to train different machine learning models like neural network and linear
regression to attain accuracy as high as possible.
We are designing the system using machine learning to help farmers in crop and
fertilizer prediction. Right crop will be recommended for a specific soil and also keeping in
mind of climatic boundaries. Also, the system provides information about the required
content and the needed amount of fertilizer, the seeds needed for planting.
With the help of our system farmers can try to grow or cultivate different varieties with
right technique, which will help farmers in maximizing their profit.
3
CHAPTER 2
LITRATURE SURVEY
1) Crop Selection Method to Maximize Crop Yield Rate using Machine Learning
Technique Authors: Rakesh Kumar, M.P. Singh, Prabhat Kumar and J.P. Singh
This paper proposed a method named Crop Selection Method (CSM) to solve crop
selection problem, and maximize net yield rate of crop over season and subsequently
achieves maximum economic growth of the country. The proposed method may improve net
yield rate of crops.
This paper contains about the research and the building of an effective agricultural
yield forecasting system based on real-time monthly weather. It is difficult to predict the
agricultural crop production because of the abnormal weather that happens every year and
rapid regional climate change due to global warming. The development of agricultural yield
forecasting system that leverages real-time weather information is urgently required. In this
research, we cover how to process the number of weather data (monthly, daily) and how to
configure the prediction system. We establish a non-parametric statistical model on the basis
of 33 years of agricultural weather information. According to the implemented model, we
predict final production using the monthly weather information. This paper contains the
results of the simulation.
4
4) Analysis of Soil Behaviour and Prediction of Crop Yield using Data Mining
Approach Monali Paul, Santosh K. Vishwakarma, Ashok Verma Computer science and
Engineering GGITS, Jabalpur
This work presents a system, which uses data mining techniques in order to predict the
category of the analyzed soil datasets. The category, thus predicted will indicate the yielding
of crops. The problem of predicting the crop yield is formalized as a classification rule, where
Naive Bayes and K-Nearest Neighbor methods are used.
5
CHAPTER 3
THEORETICAL BACKGROUND
3.1 OVERVIEW ON MACHINE LEARNING
Machine learning is an application of artificial intelligence (AI) that gives systems the
ability to automatically learn and evolve from experience without being specially
programmed by the programmer. The process of learning begins with observations or data,
such as examples, direct experience, or instruction, in order to look for patterns in data and
make better decisions in the future based on the examples that we provide. The main aim of
machine learning is to allow computers to learn automatically and adjust their actions to
improve the accuracy and usefulness of the program, without any human intervention or
assistance. Traditional writing of programs for a computer can be defined as automating the
procedures to be performed on input data in order to create output artifacts. Almost always,
they are linear, procedural and logical. A traditional program is written in a programming
language to some specification, and it has properties like:
Traditional programming works on the premise that, as long as we can define what a
program needs to do, we are confident we can define how a program can achieve that goal.
This is not always the case as sometimes, however, there are problems that you can represent
in a computer that you cannot write a traditional program to solve. Such problems resist a
procedural and logical solution. They have properties such as:
• You cannot specify how to achieve the goal of the program, only what that goal is.
• You cannot map out all the decisions the program will need to make to achieve its goal.
• You can collect only sample input data but not all possible input data for the program.
6
3.1.1 SUPERVISED LEARNING
Supervised learning takes a set of feature/label pairs, called the training set. From this
training set the system creates a generalized model of the relationship between the set of
descriptive features and the target features in the form of a program that contains a set of
rules. The objective is to use the output program produced to predict the label for a
previously unseen, unlabelled input set of features, i.e. to predict the outcome for some new
data. Data with known labels, which have not been included in the training set, are classified
by the generated model and the results are compared to the known labels. This dataset is
called the test set. The accuracy of the predictive model can then be calculated as the
proportion of the correct predictions the model labeled out of the total number of instances in
the test set.
Using programming languages to develop and implement models is more flexible and gave us
better control of the parameters to the algorithms. It also allows us to have a better
understanding of the output models produced. Some of the popular programming languages
used in the field of machine learning are:
7
• Python: Python is an extremely popular choice in the field of machine learning and AI
development. Its short and simple syntax make it extremely easy to learn
• Tensorflow: TensorFlow is an end-to-end open source platform for machine learning. It has
a comprehensive, flexible ecosystem of tools, libraries, and community resources that let
researchers pushes the state-of-the-art in ML and developers easily build and deploy ML-
powered applications. TensorFlow was originally developed by researchers and engineers
working on the Google Brain team within Google’s Machine Intelligence Research
organization to conduct machine learning and deep neural networks research. The system is
general enough to be applicable in a wide variety of other domains, as well.
3.3 SCIKIT-LEARN
SciKit learn is an open source machine learning library built for python. Since its
release in 2007, Scikit-learn has become one of the most popular open-source machine
learning libraries. Scikit-learn (also called sklearn) provides algorithms for many machine
learning tasks including classification, regression, dimensionality reduction and clustering.
The documentation for scikit-learn is comprehensive, popular and well maintained. Sklearn is
built on mature Python Libraries such as NumPy, SciPy, and matplotlib. While languages
such as R and MATLAB are extremely popular and useful for machine learning, we decided
to choose Python along with its SciKit-learn libraries as our programming language of
choice. The reasons for this are:
• We already have some familiarity and exposure to Python, and thus have a smaller learning
curve.
• Both Python and Scikit-learn have excellent documentation and tutorials available online
• The number of classic machine learning algorithms that come with Scikit-learn, and the
consistent patterns for using the different models i.e., each model can be used with the same
basic commands for setting up the data, training the model and using the model for
prediction. This makes it easier to try a range of machine learning algorithms on the same
data.
• The machine learning algorithms included with sklearn have modifiable parameters known
as hyperparameters that effect the performance of the model. These usually have sensible
default values, so that we can run them without needing a detailed knowledge or
understanding of their semantics.
Crop dataset
Modal price of crops
Standard price of crops
Soil nutrient content dataset
Rainfall Temperature dataset
• Neural Network: Neural networks are a set of algorithms, modeled loosely after the human
brain, that are designed to recognize patterns. They interpret sensory data through a kind of
machine perception, labeling or clustering raw input. The patterns they recognize are
numerical, contained in vectors, into which all real-world data, be it images, sound, text or
time series, must be translated. Neural networks help us cluster and classify. Neural Networks
are themselves general function approximations, which is why they can be applied to almost
any machine learning problem about learning a complex mapping from the input to the output
space.
Python:
For carrying out this project in the best possible manner, we decided on using Python
Language, which comes with several pre-built libraries (such as pandas, NumPy, SciPy, and
etc.) and is loaded with numerous features for implementing data science and machine
learning techniques which allowed us to design the model in the most efficient manner
possible. For building this project we utilized numerous python libraries for executing
different operations.
10
● Python - Python is a robust programming language with a wide range of capabilities. Its
broad features make working with targeted programs (including meta-programming and
meta- objects) simple. Python takes advantage of power typing as well as the integration
of reference computation and waste management waste collecting. It also supports
advanced word processing (late binding), which binds the way the words change during
the process. Patches to fewer essential sections of C Python that can give a minor
improvement in performance at an obvious price are rejected by Python developers who
try to prevent premature execution. When speed is crucial, the Python program developer
can use mod- written modules in C-languages or PyPy, a timely compiler, to submit time-
sensitive jobs.
Cython is a Python interpreter that transforms Python scripts to C and uses the C-1evel API
to call the Python interpreter directly. Python's creators attempt to make the language as fun
to use as possible. Python's architecture supports Lisp culture in terms of functionality.
Filters, maps, and job reduction, as well as a list comprehension, dictionaries, sets, and
generator expressions, are all included.
Two modules (itertools and fun tools) in the standard library use realistic Haskell and
Standard ML tools.
We're using Python because it works on a wide range of platforms. Python is a language with
no stages. Python is a as simple as English. Python have many libraries and has a simple
linguistic structure similar to English, whereas Java and C++ have complicated codes. Python
applications contain less line than programs written in other languages. That is why we
choose Python for artificial intelligence, artificial consciousness, and dealing with massive
volumes of data. Python is an article-oriented programming language. Classes, objects,
polymorphism, exemplification, legacy, and reflection are all concepts in Python .
Python has become a staple in data science, allowing data analysts and other professionals to
use the language to conduct complex statistical calculations, create data visualizations, build
machine learning algorithms, manipulate and analyze data, and complete other data-related
tasks.
11
HTML:
The webpage can be divided into multiple small sections and each section has specific
information in it. So when you write an html document you are giving browser set of
instructions on how to display content on web page
As we know that html to define the structure of the web page with no styling. To style
the webpage like font, color, size and much more we use CSS or cascading style sheets. With
help of it we can any element of our web page. To define our web page browser reads html
and CSS together.
It is of three types:
Inline
Internal
External
12
JavaScript:
JavaScript, commonly known as JS, is an open and cross platform interpreted programming
language.
It helps in creating frontend and backend applications using various frameworks. It not only
makes your website good looking but functional too.
It is a small framework in python. It has no basis for website summaries, form verification, or
any other categories where third-party libraries provide similar services. However, Flask
supports extensions that can add features to the app as if they were made in Flask itself. There
are object-related map extensions, form verification, download management, various open
authentication technologies and several tools related to the standard framework.
Flask gives the developer a variety of options when designing web applications, by giving
person with means that help them to create or build website but will not force you to rely on
or tell you what the project should look like.
13
NumPy library:
NumPy is a Python program library, which adds support for large, multi-dimensional
collections and matrices, as well as a large collection of mathematical functions developed to
work in these components.
The use of NumPy in Python is basically the same as that of MATLAB as they both translate
and allow customers to create projects faster as long as multiple tasks are focused on clusters
or networks rather than scales. Along with these critical ones, there are several options:
Pandas’ library:
It is a software library in python to decrypt and analyze data. It provides data structures and
functions to manage number tables and time series. Free software released under a three-
phase BSD license. The term is taken from the term "panel data", an econometrics term for
data sets that incorporates visibility into many identical people.
Adding or modifying data engines by a robust community that allows different applications
to be integrated into data sets. High output of a combination of data and a combination.
Hierarchical indexing provides an accurate way of dealing with large-scale data in a small
data structure.
14
Matplotlib:
John Hunter and many others built a matplotlib Python library to create graphs, charts, and
high-quality statistics. The library can change very little information about mathematics, and
it is great. Some of the key concepts and activities in matplotlib are:
Picture:
Every image is called an image, and every image is an axis. Drawing can be considered as a
way to draw multiple episodes.
Structure:
Data is the first thing that a graph should be drawn. A keyword dictionary with keys and
values such as x and y values can be declared. Next, scatter (), bar (), and pie () can be used
to create a structure and a host of other functions.
Axis:
Adjustments are possible using the number and axes obtained using the sub-sections (). Uses
a set () function to adjust x-axis and y-axis features.
15
Scikit learn:
The best Python Scikit-learn machine library. The sklearn library contains many practical
machine learning tools and mathematical modeling methods, including division, deceleration,
integration and size reduction. Machine learning models used by sklearn. Scikit-Learn
charges for tons of features and should not be used to read data or trick or summarize it.
Some of them are there to help you translate the spread.
Scikit-learn comes with many features. Some of them are here to help us explain the spread:
• Supervised learning algorithms: Consider any professional reading algorithms you may
have studied and may be part of science. Starting with the standard line models, SVM,
decision trees are all in the science toolbox. One of the main reasons for the high level of use
of scientists is the proliferation of machine learning algorithms. I started using scikit, and I
would recommend young people to learn the scikit / machine. I will solve supervised learning
problems.
• Unchecked learning algorithms: There are also a wide variety of machine learning
algorithms ranging from compilation, feature analysis, key component analysis to unchecked
neural networks.
• Contrary verification: a variety of methods are used by sky learn to ensure the accuracy
of the models followed with invisible details.
• Datasets for different toys: This was useful when studying science. I have studied SAS
for different educational data sets. It helped them a lot to support when they read the new
library.
16
CHAPTER 4
SYSTEM REQUIREMENTS SPECIFICATION
A software requirements specification (SRS) is a description of a software system to be
developed. It lays out functional and non-functional requirements, and may include a set of
use cases that describe user interactions that the software must provide. In order to fully
understand one’s project, it is very important that they come up with a SRS listing out their
requirements, how are they going to meet it and how will they complete the project. It helps
the team to save upon their time as they are able to comprehend how are going to go about
the project. Doing this also enables the team to find out about the limitations and risks early
on. Requirement is a condition or capability to which the system must conform. Requirement
Management is a systematic approach towards eliciting, organizing and documenting the
requirements of the system clearly along with the applicable attributes. The elusive
difficulties of requirements are not always obvious and can come from any number of
sources.
2. The data collected will be vectorized and sent across to the classifier.
17
4.2.1 SYSTEM CONFIGURATION
Software Configuration:
• Tools:
1. Pandas
2. Numpy
3. Tensorflow
4. Keras
5. Sickitlearn
18
CHAPTER 5
SYSTEM ANALYSIS
5.1 FEASIBILITY STUDY
Analysis is the process of finding the best solution to the problem. System analysis is
the process by which we learn about the existing problems, define objects and requirements
and evaluates the solutions. It is the way of thinking about the organization and the problem it
involves, a set of technologies that helps in solving these problems. Feasibility study plays an
important role in system analysis which gives the target for design and development.
19
5.2 ANALYSIS
20
CHAPTER 6
IMPLEMENTATION
6.1 DATA ANALYSIS
One of the first steps we perform during implementation is an analysis of the data. This
was done by us in an attempt to find the presence of any relationships between the various
attributes present in the dataset.
21
Figure 6.1.2 Crop dataset
Missing data can be dealt with in two ways. The first method is to simply remove the entire
row which contains the missing or erroneous value. While this an easy to execute method, it
is better to use only on large datasets. Using this method on small datasets can reduce the
dataset size too much, especially if there are a lot of missing values. This can severely affect
the accuracy of the result. Since ours is a relatively small dataset, we will not be using this
method.
The dataset that we used had values that were in string format so we had to transform
and encode the into integer valued so as to pass as an input to the neural network. First we
converted the data into pandas categorical data and then generated codes for crops and states
respectively we than appended these and created separated datasets. The steps are illustrated
below.
22
6.3 MACHINE LEARNING MODELS
• LOGISTIC REGRESSION MODEL
Logistic regression is a supervised learning classification algorithm used to predict the
probability of a target variable. The nature of target or dependent variable is dichotomous,
which means there would be only two possible classes.
23
• RANDOM FOREST MODEL
Random Forest has the ability to analyze crop growth related to the current climatic
condition and biophysical change. Random forest algorithm creates decision trees on
different data samples and then predicts the data from each subset and then by voting gives
better solution for the system. Random Forest uses the bagging method to train the data
which increases the accuracy of the result.
24
CHAPTER 7
RESULTS AND PERFORMANCE ANALYSIS
For the purposes of this project we have used three popular algorithms: Linear regression,
Logistic regression and Neural network. All the algorithms are based on supervised learning.
Our overall system is divided into three modules:
• Crop recommender
• Fertilizer recommender
25
• Naive Bayes
• Random Forest
26
• Decision tree
27
• Accuracy Comparison of Algorithms
28
CHAPTER 8
CONCLUSION
In this project we try to get best crop and fertilizer recommendation with the help of
machine learning. For the calculation of accuracy many machine learning techniques were
imposed or used. Numerous algorithms were used on datasets to get the best output which
leads to best crop and fertilizer recommendation for particular soil of particular region. This
system will help farmers to visualize crop yields based on that climatic and subsistence
boundaries. Using this farmer can decide whether to plant that crop or to look for another
crop if yield forecasts are incorrect. This tool can help the farmer to make the best decisions
when it comes to growing something harvest. It may also predict the negative effects of the
plant. Currently our farmers use outdated technology or not use effectively, so there can be an
opportunity of the wrong choice of cultivated crops that will reduce the profit by production.
To reduce these types of loss we try to create a farmer-friendly system, which will help in
predicting which crop is best for a specific soil and this project will give the recommendation
about the fertilizer needed by the soil for cultivation, seeds needed for cultivation,
expectations yield and market price. Thus, this enables farmers to make the right choice in
choosing a crop farming so that the agricultural sector can develop with new ideas.
Current Market Conditions and analysis for information on crop market rates,
production costs, fertilizer.
The mobile app can be developed to assist farmers with uploading farm photos.
29
APPENDECIES
APPENDIX - I
SAMPLE CODING
# Importing essential libraries and modules
import numpy as np
import pandas as pd
import requests
import config
import pickle
import io
import torch
disease_classes = ['Apple___Apple_scab',
'Apple___Black_rot',
'Apple___Cedar_apple_rust',
'Apple___healthy',
'Blueberry___healthy',
30
'Cherry_(including_sour)___Powdery_mildew',
'Cherry_(including_sour)___healthy',
'Corn_(maize)___Cercospora_leaf_spot Gray_leaf_spot',
'Corn_(maize)___Common_rust_',
'Corn_(maize)___Northern_Leaf_Blight',
'Corn_(maize)___healthy',
'Grape___Black_rot',
'Grape___Esca_(Black_Measles)',
'Grape___Leaf_blight_(Isariopsis_Leaf_Spot)',
'Grape___healthy',
'Orange___Haunglongbing_(Citrus_greening)',
'Peach___Bacterial_spot',
'Peach___healthy',
'Pepper,_bell___Bacterial_spot',
'Pepper,_bell___healthy',
'Potato___Early_blight',
'Potato___Late_blight',
'Potato___healthy',
'Raspberry___healthy',
'Soybean___healthy',
'Squash___Powdery_mildew',
'Strawberry___Leaf_scorch',
'Strawberry___healthy',
'Tomato___Bacterial_spot',
'Tomato___Early_blight',
'Tomato___Late_blight',
31
'Tomato___Leaf_Mold',
'Tomato___Septoria_leaf_spot',
'Tomato___Spider_mites Two-spotted_spider_mite',
'Tomato___Target_Spot',
'Tomato___Tomato_Yellow_Leaf_Curl_Virus',
'Tomato___Tomato_mosaic_virus',
'Tomato___healthy']
disease_model_path = 'models/plant_disease_model.pth'
disease_model.load_state_dict(torch.load(
disease_model_path, map_location=torch.device('cpu')))
disease_model.eval()
crop_recommendation_model_path = 'models/RandomForest.pkl'
crop_recommendation_model = pickle.load(
open(crop_recommendation_model_path, 'rb'))
def weather_fetch(city_name):
"""
:params: city_name
"""
32
api_key = config.weather_api_key
base_url = "http://api.openweathermap.org/data/2.5/weather?"
response = requests.get(complete_url)
x = response.json()
if x["cod"] != "404":
y = x["main"]
humidity = y["humidity"]
else:
return None
"""
:params: image
"""
transform = transforms.Compose([
transforms.Resize(256),
transforms.ToTensor(),
33
])
image = Image.open(io.BytesIO(img))
img_t = transform(image)
img_u = torch.unsqueeze(img_t, 0)
yb = model(img_u)
prediction = disease_classes[preds[0].item()]
return prediction
# FLASK APP
app = Flask(__name__)
@ app.route('/')
def home():
@ app.route('/crop-recommend')
def crop_recommend():
34
title = '- Crop Recommendation'
@ app.route('/fertilizer')
def fertilizer_recommendation():
@ app.route('/crop-predict', methods=['POST'])
def crop_prediction():
if request.method == 'POST':
N = int(request.form['nitrogen'])
P = int(request.form['phosphorous'])
K = int(request.form['pottasium'])
ph = float(request.form['ph'])
rainfall = float(request.form['rainfall'])
# state = request.form.get("stt")
city = request.form.get("city")
35
if weather_fetch(city) != None:
my_prediction = crop_recommendation_model.predict(data)
final_prediction = my_prediction[0]
else:
@ app.route('/fertilizer-predict', methods=['POST'])
def fert_recommend():
crop_name = str(request.form['cropname'])
N = int(request.form['nitrogen'])
P = int(request.form['phosphorous'])
K = int(request.form['pottasium'])
# ph = float(request.form['ph'])
df = pd.read_csv('Data/fertilizer.csv')
36
nr = df[df['Crop'] == crop_name]['N'].iloc[0]
pr = df[df['Crop'] == crop_name]['P'].iloc[0]
kr = df[df['Crop'] == crop_name]['K'].iloc[0]
n = nr - N
p = pr - P
k = kr - K
max_value = temp[max(temp.keys())]
if max_value == "N":
if n < 0:
key = 'NHigh'
else:
key = "Nlow"
if p < 0:
key = 'PHigh'
else:
key = "Plow"
else:
if k < 0:
key = 'KHigh'
else:
key = "Klow"
37
response = Markup(str(fertilizer_dic[key]))
def disease_prediction():
if request.method == 'POST':
return redirect(request.url)
file = request.files.get('file')
if not file:
try:
img = file.read()
prediction = predict_image(img)
prediction = Markup(str(disease_dic[prediction]))
except:
pass
if __name__ == '__main__':
app.run(debug=False)
38
APPENDIX - II
OUTPUT SCREESHOTS
Description: The about us page is often a reflection of the purpose and personality of the
business and its owners or top employees
39
Figure 9.3 Our services
Description: Our service is a simple ML and DL based website which recommends the best
crop to grow, fertilizers to use and the diseases caught by your crops
crops.
40
CROP RECOMMEDATION SYSTEM
41
FERTILIAZER RECOMMEDATION SYSTEM
Description: This recommends the fertilizer to your crop. Based on your input
i data.
42
DISEASE PREDICTION
43
REFERRENCE
[1] Rakesh Kumar, M.P. Singh, Prabhat Kumar and J.P. Singh, “Crop Selection Method to
Maximize Crop Yield Rate using Machine Learning Technique”, International Conference on
Smart Technologies and Management for Computing, Communication, Controls, Energy and
Materials, 2015
[2] Haedong Lee and Aekyung Moon, “Development of Yield Prediction System Based on
Real-time Agricultural Meteorological Information”, 16th International Conference on
Advanced Communication Technology, 2014
+[3] T.R. Lekhaa, “Efficient Crop Yield and Pesticide Prediction for Improving Agricultural
Economy using Data Mining Techniques”, International Journal of Modern Trends in
Engineering and Science (IJMTES), 2016, Volume 03, Issue 10
[4] Jay Gholap, Anurag Ingole, Jayesh Gohil, Shailesh Gargade and Vahida Attar, “Soil Data
Analysis Using Classification Techniques and Soil Attribute Prediction”, International
Journal of Computer Science Issues, Volume 9, Issue 3
[5] Anshal Savla, Parul Dhawan, Himtanaya Bhadada, Nivedita Israni, Alisha Mandholia,
Sanya Bhardwaj (2015), ‘Survey of classification algorithms for formulating yield prediction
accuracy in precision agriculture’, Innovations in Information, Embedded and
Communication systems (ICIIECS).
[8] Tom M. Mitchell, Machine Learning, India Edition 2013, McGraw Hill Educa- tion.
[9] https://data.gov.in/g
[10] Kagglehttps://www.kaggle.com/notebook
44