0% found this document useful (0 votes)
21 views39 pages

LL LL LLLLL LLLLL

Uploaded by

gawoy90443
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
21 views39 pages

LL LL LLLLL LLLLL

Uploaded by

gawoy90443
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PPTX, PDF, TXT or read online on Scribd
You are on page 1/ 39

UNIT 2:

AI PROJECT CYCLE
Unit 2 : INTRODUCTION TO AI PROJECT
CYCLE

❖SESSION 1 :Introduction to AI Project


❖SESSION 2: Understanding Problem Scoping
and SDG
❖SESSION 3: Simplifying Data Acquisition
❖SESSION 4: Data Visualization
❖SESSION 5: Introduction to modelling
❖SESSION 6: Evaluating the idea
What is Project Cycle ?

A Project Cycle is a sequence of tasks that must be completed


to attain a certain goal.It has a clear start and end date and
passes through several phases of development.
Example : Creating a birthday card.
1. Checking the factors like budget,etc Which will help us decide the next steps
and understanding the Project.
2. Acquiring data from different sources like online, with friends etc for Designs
and ideas.
3. Making a list of the gathered data / analysing the data.
4. Creating or Modelling a card on the basis of the data collected.
5. Showing it to Parents or cousins to let them check it or evaluate it.
Components of Project Cycle
I Problem Scoping
Problem Scoping refers to understanding a problem,finding
out various factors which affect the problem, define the goal
or aim of the project.
• Identifying a problem and having a vision to solve it.
• Determining the nature ,boundaries and different aspects of a
problem.
• A lot of times we are unable to observe any problem in our
surroundings. In that case, we can take a look at the Sustainable
Development Goals.
• 17 goals have been announced by the United nations which are
termed as the Sustainable Development Goals.
Sustainable Development: To Develop for the present without
exploiting the resources of the future.

The Sustainable Development Goals (SDGs),also known as the


Global Goals, were adopted by all United Nations Member States
in 2015 as a universal call to action to end poverty, protect the
planet and ensure that all people enjoy peace and prosperity.

● 17 goals announced by United Nations.


● Aim to achieve them by 2030.
● Pledge taken by all the member nations of the UN.
We use the 4Ws Problem Canvas to have a
deeper understanding around the problem
1. Who - “Who” part helps us in comprehending and categorizing who all
are affected directly and indirectly with the problem and who are called
the Stakeholders (Who are the stakeholders? What do you know about
them?)

2. What - “What” part helps us in understanding and identifying the nature


of the problem and under this block, we also gather evidence to prove
that the problem we have selected exists.(What is the problem?How do
you know it is a problem?)

3. Where- "Where” does the problem arise, situation,context,and location.


(What is the context/situation in which the stakeholders experience the
problem?)

4. Why - “Why” is the given problem worth solving.(What would be of key


value to the stakeholders? How would it improve the situation?)
• many of the restaurants arrange for
buffets to offer a variety of food items to
their customers.
• prepares food in bulk as they expect a
good crowd to come and enjoy their
food.
• But in most cases, after the day ends, a
lot of food is left which becomes
unusable for the restaurant as they do
not wish to serve stale food to their
customers the next day.
• So, every day, they prepare food in
large quantities keeping in mind the
probable number of customers walking
into their outlet. But if the expectations
are not met, a good amount of food gets
wasted which eventually becomes a
loss for the restaurant as they either
have to dump it or give it to hungry
people for free.
• And if this daily loss is taken into
account for a year, it becomes quite a
big amount.
Problem Statement Template
The Problem statement template leads us towards the goal of our
project which can now be stated as:
“To be able to predict the quantity of food dishes to be
prepared for everyday consumption in restaurant buffets.”
WHAT IS DATA ACQUISITION?
Data Acquisition consists of two words:

1. Data : Data refers to the raw facts , figures, or piece of facts, or


statistics collected for reference or analysis.

2. Acquisition : Acquisition refers to acquiring data for the project.

Whenever we want an AI project to be able to predict an output, we need


to train it first using data.The stage of acquiring data from the relevant
sources is known as data acquisition.
DATA FEATURES
Data features refer to the type of data you want to collect. For any AI project to be efficient, the
training data should be authentic and relevant to the problem statement scoped.

Here two terms are associated with this:


1. Training Data: The collected data through the system is known as training data. In other
words the input given by the user in the system can be considered as training data.
2. Testing Data: The result data set or processed data is known as testing data. In other words,
the output of the data is known as testing data

Example, If you want to make an AI system which can predict the salary of any employee
based on his previous salaries, you would feed the data of his previous salaries into the
machine. This is the data with which the machine can be trained. Now, once it is ready, it will
predict his next salary efficiently. The previous salary data here is known as Training Data
while the next salary prediction data set is known as the Testing Data.
Data sources
• Various ways to collect data are
• find a reliable source of data from where
some authentic information can be taken.
• data can be open-sourced and not
someone’s property.Extracting private
data can be an offence.
• One of the most reliable and authentic sources of information,
are the open-sourced websites hosted by the government.
1. Web Scraping means collecting data from web using some technologies We
use it for monitoring prices, news and etc Example Web Scrapping
2. Sensors are very Important but very simple to understand. Sensors are the part
of IoT Internet of things which collect the physical data and detect the changes
3. Camera captures the visual information and then that information which is called
image is used as a source of data. Cameras are used to capture raw visual data
4. Observations: When we observe something carefully we get some information
For ex Scientists Observe creatures to study them Observations is a time
consuming data source.
5. API :Application Programming interface API is a messenger which takes
requests and tells the system about requests and gives the response
Ex Twitter API, Google Search API
6. Surveys: The survey is a method of gathering specific information from a sample
of people Example a census survey for analyzing the population
TYPES OF DATA
The structure classification is divided into 3 categories:

1. Structured Data : It can have a specific pattern or set of rules. These data have a
simple structure and stores the data in specific forms such as tabular form.
Example, The cricket scoreboard, Your school time table, Exam datasheet etc.
2. Unstructured Data : The data structure which doesn't have any specific pattern
or constraints as well as can be stored in any form is known as unstructured data.
Mostly the data that exists in the world is unstructured data. Example, Youtube
Videos, Facebook Photos, Dashboard data of any reporting tool etc.
3. Semi-Structured Data : It is the combination of both structured and unstructured
data. Some data can have a structure like a database whereas some data can
have markers and tags to identify the structure of data.
III DATA EXPLORATION
• To analyse the data,
you need to visualise it
in some user-friendly
format so that you can:
quickly get a sense of
the trends,
relationships and
patterns contained
within the data.

• To visualise data, we
can use various types
of visual
Data Exploration refers to the techniques and tools used to
visualize data through complex statistical methods.
Advantages of Data Visualization
❖A better understanding and provides insights into data
❖Allows user interaction
❖Provide real time analysis
❖Help to make decisions
❖Reduces complexity of data
❖Provides the relationships and patterns contained within
data
❖Define a strategy for your data model
❖Provides an effective way of communication among users
How to select a proper graph?
1.Comparison of Values -Show periodical changes.
ie Bar Chart
2. Comparison of Trends -Show changes over a period of
time
ie Line Chart
3. Distribution of Data according to categories-Show data
according to category ie Histogram
4.Highlight a portion of a whole Highlight data according to
value ie Pie Chart
5.Show the relationship between data -Multiple charts can be
IV MODELING
AI Modelling refers to developing algorithms, also
called models which can be trained to get intelligent
outputs or writing codes to make a machine artificially
intelligent
● The graphical representation makes the data
understandable for humans as we can discover
trends and patterns out of it
•But when it comes to machine accessing and
analysing data, it needs the data in the most basic
form of numbers (which is binary 0 s and 1 s) and
when it comes to discovering patterns and trends in
data, the machine goes for mathematical
representations of the same
Rule Based Approach

• Rule Based Approach


refers to the AI
modelling where the
relationship or patterns
in data are defined by
the developer

• The machine follows


the rules or
instructions mentioned
by the developer and
performs its task
accordingly
LEARNING BASED APPROACH
● AI modelling where the relationship or patterns in data are
not defined by the developer
● In this approach, random data is fed to the machine and it
is left on the machine to figure out patterns and trends out
of it
● Generally this approach is followed when the data is
unlabeled and too random for a human to make sense out
of it Thus, the machine looks at the data, tries to extract
similar features out of it and clusters same datasets
together
● In the end as output, the machine tells us about the trends
SUPERVISED LEARNING
• In a supervised learning model, the dataset which is
fed to the machine is labelled.
• the dataset is known to the person who is training the
machine only then he/she is able to label the data.
• A label is some information which can be used as a
tag for data.
• For example, students get grades according to the
marks they secure in examinations. These grades are
labels which categorise the students according to their
marks.
There are two types of Supervised Learning models:
Classification: Where the
data is classified according to
the labels. For example, in
the grading system, students
are classified on the basis of
the grades they obtain with
respect to their marks in the
examination. This model
works on discrete dataset
which means the data need
not be continuous.
Regression: Such
models work on
continuous data. For
example, if you wish to
predict your next salary,
then you would put in the
data of your previous
salary, any increments,
etc., and would train the
model. Here, the data
which has been fed to the
machine is continuous.
UNSUPERVISED LEARNING
Unsupervised learning models can be further divided into
two categories:
Clustering: Refers to the
unsupervised learning algorithm
which can cluster the unknown
data according to the patterns or
trends identified out of it. The
patterns observed might be the
ones which are known to the
developer or it might even come
up with some unique patterns
out of it.
Dimensionality reduction
• Dimensionality reduction is the process of reducing the number
of features (or dimensions) in a dataset while retaining as much
information as possible. This can be done for a variety of
reasons, such as to reduce the complexity of a model, to
improve the performance of a learning algorithm, or to make it
easier to visualize the data.
• As we reduce the dimension of an entity, the information which it
contains starts getting distorted. For example, if we have a ball in
our hand, it is 3-Dimensions right now. But if we click its picture,
the data transforms to 2-D as an image is a 2-Dimensional entity.
Now, as soon as we reduce one dimension, at least 50% of the
information is lost as now we will not know about the back of the
ball. Whether the ball was of same colour at the back or not? Or
was it just a hemisphere? If we reduce the dimensions further,
more and more information will get lost.
• Hence, to reduce the dimensions and still be able to make sense
out of the data, we use Dimensionality Reduction.
EVALUATION
Once a model has been made and trained, it needs to
go through proper testing so that one can calculate the
efficiency and performance of the model. Hence, the
model is tested with the help of Testing Data (which
was separated out of the acquired dataset at Data
Acquisition stage) and the efficiency of the model is
calculated on the basis of the parameters mentioned
below:

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy