Ai ML Unit 1
Ai ML Unit 1
Definitions of AI vary
Artificial Intelligence is the studyof
systems that
think like humansthink rationally
4
WHATISAI?
A systemisrational if itdoes the“right thing,” givenwhat itknows.
1)Actinghumanly:TheTuringTestapproach
yieldcorrect conclusionswhengivencorrectpremises
1.Philosophy
2.Mathematics
3.Economics
4.Neuroscience
5.Psychology
6.Computerengineering
7.Controltheoryandcybernetics
8.Linguistics
1.Philosophy
Materialism which holds that the brain’s operation accordingto the lawsof physics
constitutes the mind.
Thomas Bayes (1702–1761),proposed arule for updating probabilitiesin the lightof new
evidence.
3.Economics
How should we makedecisions so asto maximize payoff?
•How should we dothis when othersmay not goalong?
•How should we dothis when thepayoff may befar in the future?
4.Neuroscience
How do brains processinformation?
Figure 1.3 shows that computers have a cycle timethat is amillion times faster than
a brain.
5.Psychology
How do humans and animalsthink and act?
6.Computerengineering
8.Linguistics
2)Thebirthofartificialintelligence(1956)
3)Earlyenthusiasm,greatexpectations(1952–1969)
4)Adoseofreality(1966–1973)
5)Knowledge-basedsystems:Thekeytopower?(1969–1979)
6)AIbecomesanindustry(1980–present)
7)Thereturnofneuralnetworks(1986–present)
8)AIbecomesascientific(1987–present)
9)Theemergenceofintelligentagents(1995–present)
THESTATEOFTHEART
fewapplicationsofAI:
1)Autonomousplanningandscheduling:
2.Gameplaying:
3.Autonomouscontrol:
4.Diagnosis:
5.LogisticsPlanning:
6.Robotics:
7.Languageunderstandingandproblemsolving:
1)Autonomousplanningandscheduling:
2)Gameplaying:
4.Diagnosis:
7.Languageunderstandingandproblemsolving:
PROVERB (Littmanetal.,1999) is a computer program that solves
crossword puzzles better than most humans, usingconstraints on
possible word fillers, a large database of past puzzles, and a variety of
informationsources including dictionaries and online databases such as
a list of movies and the actors thatappear inthem.
AGENTSANDENVIRONMENTS
Anagentis anything that can be viewed as perceiving
itsenvironmentthrough sensorsand actingupon that
environmentthrough actuators.
A human agent has eyes, ears, and otherorgansforsensorsand
hands, legs, mouth, and otherbodypartsforactuators.
1.rationalagent
2.Performancemeasures
3.Rationality
4.Omniscience,learning,andautonomy
Arationalagentis one thatdoes the rightthing—conceptuallyspeaking
everyentry inthetablefor the agentfunction is filledout correctly.
The right action is the one that will cause the agent tobe most
successful.
Ruleforperformancemeasure
Asageneralrule,itisbettertodesignperformancemeasuresaccording
towhat one actually wants inthe
environment,ratherthanaccordingtohow one thinks the agent
should behave.
Rationality
Foreachpossibleperceptsequence,arationalagentshouldselectanact
ionthatisexpectedtomaximizeitsperformancemeasure,giventheevid
enceprovidedbytheperceptsequenceandwhateverbuilt-
inknowledgetheagenthas.
The performance measure awards one pointfor each clean square
at each timestep, over a "lifetime" of 1000 time steps.
1.Specifyingthetaskenvironment
2.Propertiesoftaskenvironments
a)Fullyobservablevs.partiallyobservable.
b)Deterministicvs. stochastic.
c)Episodicvssequential
d)Staticvsdynamic.
e)Discretevscontinuous.
f)Singleagent vsmultiagent.
1.Specifyingthetaskenvironment
group alltogether under the heading of thetaskenvironmentcallthisas
PEAS(Performance,Environment,Actuators,Sensors)description.
PEASelementsforanumberofadditionalagenttypes
2.Propertiesoftaskenvironments
a)Fullyobservablevs.partiallyobservable.
Fullyobservable
Ifan agent's sensors give itaccesstothecompletestateoftheenvironmentat each point
in time, then we say that thetask environment is fullyobservable.
partiallyobservable
In episodic environments,
thechoiceofactionineachepisodedependsonlyontheepisodeitse
lf.
Sequential
In sequential environments, on the other hand, thecurrent
decisioncould affect all future decisions.
Chess, whenplayedwithaclock,issemidynamic.
Taxidriving is clearlydynamic
Taxidrivingis a continuousstateandcontinuous-timeproblem:
Singleagentvsmultiagent.
agent=architecture+program
TypesofAgents:-
1)Simplereflexagents;
2)Model-basedreflexagents;
3)Goal-basedagents;and
4)Utility-basedagents.
ifcar-in-front-is-brakingtheninitiate-braking.
The agent in Figure 2.10 will work onlyifthecorrectdecision
canbemadeonthebasisofonlythecurrentpercept-
thatis,onlyiftheenvironrnenitsfullyobservable.
2)Model-basedreflexagents
The most effective way to handlepartialobservabilityis for the agent
to keeptrackofthepartoftheworlditcan'tseenow.That is, the agent should
maintain some sort of internalstatethat depends on thepercept history
and thereby reflects at least some of the unobserved aspects of the
current state.
Figure2.11Amodel-basedreflexagent.
function UPDATE-STATE, which is responsible for creating the new
internal state description.
Goal-basedagents
Forexample, at a road junction, the taxi can turn left, turn right, or
go straight on.
The correct decision depends on where the taxi is trying to get to. In
other words, aswell as a current state description, the agent needs
some sort of goal information thatdescribes situations that are
desirable-for example, being at thepassenger'sdestination.
Agentkeepstrackoftheworldstateaswellassetofgoalsit’stryingto
achieve:chooses
actionsthatwill(eventually)leadtothegoal(s).
•Key difference wrt Model-Based Agents:
For example, there are many action sequences that will get the taxi
to its destination (there by achievingthe goal) but some are
quicker, safer, more reliable, or cheaper thanothers.
Whentherearemultiplepossiblealternatives,howtodecidewhic
honeisbest?
Usedecisiontheoreticmodels:e.g.,fastervs.safer.
Learningagents
Figure2.15Ageneralmodeloflearningagents.
Learning Agents
Four conceptual components
Learning element
Making improvement
Performance element
Selecting external actions
Critic
Tells the Learning element howwell the agentis doing with
respect to fixed performancestandard.
(Feedback from user orexamples, good ornot?)
Problem generator
Suggest actions that will leadto new andinformative
experiences.
Adaptandimproveovertime
theyhavetheabilitytoimproveperformancethroughlearning
The learning elementuses feedback from the criticon how the agent is doing
and determines how theperformance elementshould bemodified to do better
in the future.
The critic tells the learning element how well the agent is doing with
respect to a fixed performancestandard.
The point isthat ifthe performance element had its way, it would keep
doing the actions that are best,given what it knows.