Planning and Machine Learning
Planning and Machine Learning
Planning System
For any planning system, we need the domain description, action specification, and
goal description. A plan is assumed to be a sequence of actions and each action has
its own set of preconditions to be satisfied before performing the action and also some
effects which can be positive or negative.
So, we have Forward State Space Planning (FSSP) and Backward State Space Planning
(BSSP) at the basic level.
1. Forward State Space Planning (FSSP)
FSSP behaves in a similar fashion like forward state space search. It says that given a
start state S in any domain, we perform certain actions required and acquire a new
state S’ (which includes some new conditions as well) which is called progress and this
proceeds until we reach the goal state. The actions have to be applicable in this case.
Disadvantage: Large branching factor
Advantage: Algorithm is Sound
2. Backward State Space Planning (BSSP)
BSSP behaves in a similar fashion like backward state space search. In this, we move
from the goal state g towards sub-goal g’ that is finding the previous action to be done
to achieve that respective goal. This process is called regression (moving back to the
previous goal or sub-goal). These sub-goals have to be checked for consistency as
well. The actions have to be relevant in this case.
Disadvantage: Not a sound algorithm (sometimes inconsistency can be found)
Advantage: Small branching factor (very small compared to FSSP)
PLANNING
Planning refers to the process of computing several steps of a problem solving before
executing any of them.Planning is useful as a problem solving technique for non
decomposable problem.
Components of Planning System:
In any general problem solving systems, elementary techniques to perform following
functions are required
Choose the best rule (based on heuristics) to be applied
Apply the chosen rule to get new problem state
Detect when a solution has been found
Detect dead ends so that new directions are explored.
To choose the rules ,
first isolate a set of differences between the desired goal state and current state,
PLANNING AND MACHINE LEARNING
Initial State
Armempty
clear(block2)
ontable(block2)
ontable(block1)
clear(block1)
Goal State
Armempty
ontable(block2)
on(block1, block2)
clear(block1)
We have to generate a plan to reach goal state from initial state given.In this example
the initial state has two blocks Block1 and Block 2.Both is placed on table.To reach the
goal state first we have to
PLANNING AND MACHINE LEARNING
PICKUP(Block 1)
We need to check whether we reach goal state or not ,after completion of each and
every operation.Here the environment looks like,
Hold(block1)
Clear(Block2)
OnTable(Block2)
This is not the goal state .so ,we have to continue the process. Next the block 1 needs
to be place on block 2,to achieve this do the operation STACK(Block1,Block2). After
this operation the environment looks like,
PLANNING AND MACHINE LEARNING
ArmEmpty,on(Block1,Block2),Clear(Block1),OnTable(Block2)
K(Agent1, K(Agent2, On(A,B) ) ], means Agent1 knows that Agent1 knows that A is on
B.
K(Agent1, On(A,B)) V K(Agent1, On(A,C) ) means that either Agent1 knows that A is on
B or it knows that A is on C.
K(Agent1, On(A,B)) V K(Agent1, ¬On(A,B) ) means that either Agent1 knows whether
or not A is on B.
Advertisement
Knowledge Axioms:
The operators ∧ and V have compositional semantics (depends on truth value) , but
the semantics of K are not compositional. The truth value of K(Agent1, On(A,B) ) for
example, cannot necessarily be determined from the properties of K, the denotation of
Agent1 and the truth value of On(A,B). K Operator is said to be referentially opaque.
Example in Planning Speech Action:
We can treat speech acts just like other agent systems. Our agent can use a plan-
generating system to make plans comprising speech acts and other actions. To do so,
it needs a model of the effects of these actions.
Consider for example, Tell( A, φ ) , where A is Agent and φ is true.
We could model the effects of that action by the STRIPS rule :
Tell( A, φ ) :
Precondition : Next_to(A) ∧ φ ∧ ¬K(A, φ)
Delete : ¬K(A, φ)
Add : K(A, φ)
The precondition Next_to(A) ensures that our agent is close to agent A to enable
communication.
The precondition φ is imposed to ensure that our agent actually believes φ before it
can inform another agent about the truth.
The precondition ¬K(A, φ) ensure that our agent does not communicate redundant
information.
Machine Learning
Machine learning is a subset of AI, which uses algorithms that learn from data to make
predictions. These predictions can be generated through supervised learning, where
algorithms learn patterns from existing data, or unsupervised learning, where they
discover general patterns in data.
The Importance of Machine Learning
Here are some reasons why it’s so essential in the modern world:
Data processing. One of the primary reasons machine learning is so important is its
ability to handle and make sense of large volumes of data. With the explosion of
digital data from social media, sensors, and other sources, traditional data analysis
PLANNING AND MACHINE LEARNING
methods have become inadequate. Machine learning algorithms can process these
vast amounts of data, uncover hidden patterns, and provide valuable insights that can
drive decision-making.
Driving innovation. Machine learning is driving innovation and efficiency across
various sectors. Here are a few examples:
Healthcare. Algorithms are used to predict disease outbreaks, personalize patient
treatment plans, and improve medical imaging accuracy.
Finance. Machine learning is used for credit scoring, algorithmic trading, and fraud
detection.
Retail. Recommendation systems, supply chains, and customer service can all benefit
from machine learning.
The techniques used also find applications in sectors as diverse as agriculture,
education, and entertainment.
Enabling automation. Machine learning is a key enabler of automation. By learning
from data and improving over time, machine learning algorithms can perform
previously manual tasks, freeing humans to focus on more complex and creative
tasks. This not only increases efficiency but also opens up new possibilities for
innovation.
Machine learning can be broadly classified into three types based on the nature of the
learning system and the data available: supervised learning, unsupervised learning,
and reinforcement learning. Let's delve into each of these:
Supervised learning
Supervised learning is the most common type of machine learning. In this approach,
the model is trained on a labeled dataset. In other words, the data is accompanied by
a label that the model is trying to predict. This could be anything from a category
label to a real-valued number.
The model learns a mapping between the input (features) and the output (label)
during the training process. Once trained, the model can predict the output for new,
unseen data.
Unsupervised learning
PLANNING AND MACHINE LEARNING
This type of learning is often used for clustering and dimensionality reduction.
Clustering involves grouping similar data points together, while dimensionality
reduction involves reducing the number of random variables under consideration by
obtaining a set of principal variables.
Reinforcement learning
Adaptive Learning
The fourth generation of machine intelligence, adaptive learning, creates the first truly
integrated human and machine learning environment. For text analytics, this has
given us the most accurate analytics to date, allowing us to get actionable information
in many areas for the first time. In the examples we will share here, we show that
adaptive learning is 95% accurate in predicting people’s intention to purchase a car.
Adaptive learning correlates with actual sales, unlike any previous approach to
Machine Intelligence.
Adaptive learning combines the previous generations of rule-based, simple machine
learning, and deep learning approaches to machine intelligence. Human analysts are
optimally engaged in making the machine intelligence smarter, faster, and easier to
interpret, building on a network of the previous generations of machine intelligence.
The first generation of machine intelligence meant that people manually created
rules. For example, in text analytics someone might create a rule that the word “Ford”
followed by “Focus” meant that “Ford” referred to a car, and they would create a
separate rule that “Ford” preceded by “Harrison” meant that “Ford” referred to a
person.
The rule-based approach is very time consuming and not very accurate. Even after an
analyst has exhausted all the words and phrases they can think of, there are always
other contexts and new innovations that aren’t captured. For one of our clients, their
experts analysts were only able to capture 11% of the documents they wanted to
analyze using rules: this clearly is too limited.
The dominant form of machine intelligence today is simple machine learning. Simple
machine learning uses statistical methods to make decisions about data processing.
For example, a sentence might have the word “Ford” labeled as a car, and the
machine learning algorithm will learn by itself that the following word “Focus” is
evidence that “Ford” is a car in this context.
Simple machine learning can be fast, provided that you already have labeled
examples for ‘supervised learning’. It also tends to be more accurate, because
statistics are usually better than human intuition in deciding which features (like
words and phrases) matter. The major drawback for supervised machine learning is
that you need the labeled examples: if you have too few labels or the labels aren’t
PLANNING AND MACHINE LEARNING
representative of the entire data set, then the accuracy is low or limited to a specific
domain.
There has been a recent rise in the use of machine learning that learns more
sophisticated relationships between features, known as deep learning. For example, if
you had the sentence “We Will Let Harrison Ford Focus on Star Wars”, there is
conflicting evidence between “Harrison” and “Focus” about whether “Ford” is a person
or a car.
Deep learning can automatically learn how to use combinations of features when
making a decision. For simple machine learning, a human has to tell the algorithm
which combination of features to consider. Deep learning often cuts down on the
amount of human time needed and typically gets up to 5% more accurate results than
simple machine learning for text analytics–although only when applied to data from
the same sources as it learned from.
Adaptive learning brings human analysts into the process at every step. This is in
contrast to rule-based, simple machine learning and deep learning approaches, where
the humans only create rules and label data at the start of the process. For example, if
you had the sentence “We Will Help Tom Ford Escape from New York”, and your
system hadn’t seen any examples of “Tom Ford” or “Ford Escape”, you will need
human input to build the knowledge.
Adaptive learning systems require the least human effort because they only require
human input when it matters most and continually expand their knowledge when new
information is encountered. As we show here, they are also the most accurate. They
combine the three other types of machine intelligence, adding new types of
‘unsupervised machine learning’ and methods for optimizing the input from multiple,
possibly disagreeing, humans.
Explanation-Based Learning
Explanation-based learning in artificial intelligence is a problem-solving method that
involves agent learning by analyzing specific situations and connecting them to
previously acquired information. Also, the agent applies what he has learned to solve
similar issues. Rather than relying solely on statistical analysis, EBL algorithms
incorporate logical reasoning and domain knowledge to make predictions and identify
patterns.
Key Characteristics of EBL
1. Use of Domain Knowledge: EBL relies heavily on pre-existing domain
knowledge to explain why a particular example is a valid instance of a concept.
This knowledge helps the system to generalize the learned concept to new,
similar situations.
2. Focused Learning: EBL focuses on understanding the essential features of an
example that are necessary to achieve a goal or solve a problem. This contrasts
with other learning methods that may treat all features equally or rely on
statistical correlations.
PLANNING AND MACHINE LEARNING
3. Efficiency: Since EBL can learn from a single example by generalizing from it, it
is computationally efficient compared to other learning methods that require
large datasets for training.
How Explanation-Based Learning Works?
Explanation-Based Learning follows a systematic process that involves the following
steps:
1. Input Example: The learning process begins with a single example that the
system needs to learn from. This example is typically a positive instance of a
concept that the system needs to understand.
2. Domain Knowledge: The system uses domain knowledge, which includes
rules, concepts, and relationships relevant to the problem domain. This
knowledge is crucial for explaining why the example is valid.
3. Explanation Generation: The system generates an explanation for why the
example satisfies the concept. This involves identifying the relevant features
and their relationships that make the example a valid instance.
4. Generalization: Once the explanation is generated, the system generalizes it
to form a broader concept that can apply to other similar examples. This
generalization is typically in the form of a rule or a set of rules that describe the
concept.
5. Learning Outcome: The outcome of EBL is a generalized rule or concept that
can be applied to new situations. The system can now use this rule to identify or
solve similar problems in the future.
Example of Explanation-Based Learning in AI
Scenario: Diagnosing a Faulty Component in a Car Engine
Context: Imagine you have an AI system designed to diagnose problems in car
engines. One day, the system is given a specific example where the engine fails to
start. After analyzing the case, the system learns that the failure was due to a faulty
ignition coil.
Step 1: Input Example
The system is provided with a scenario where a car engine fails to start. The
diagnostic information indicates that the cause is a faulty ignition coil.
Step 2: Use of Domain Knowledge
The AI system has pre-existing domain knowledge about car engines. It knows how
the ignition system works, the role of the ignition coil, and the conditions under which
an engine would fail to start.
Step 3: Explanation Generation
Using this domain knowledge, the system generates an explanation for why the
engine failure occurred:
PLANNING AND MACHINE LEARNING
Ignition System Knowledge: The system understands that the ignition coil is
responsible for converting the battery’s low voltage to the high voltage needed
to create a spark in the spark plugs.
Faulty Coil Impact: It explains that if the ignition coil is faulty, it will fail to
generate the necessary high voltage, resulting in no spark, which prevents the
engine from starting.
Step 4: Generalization
The system then generalizes this explanation to form a rule:
General Rule: “If the engine fails to start and the ignition coil is faulty, then the
cause of the failure is likely due to the ignition coil not providing the necessary
voltage to the spark plugs.”
Step 5: Learning Outcome
The AI system has now learned a new diagnostic rule that can be applied to future
cases:
Future Application: In future diagnostics, if the system encounters a similar
scenario where the engine fails to start, it can use this learned rule to quickly
check the ignition coil as a potential cause.
Applications of Explanation-Based Learning
Explanation-Based Learning is particularly useful in domains where understanding the
reasoning behind decisions is critical.
Some of the notable applications of EBL include:
Medical Diagnosis: EBL can be used in medical diagnosis systems to learn
from specific cases and generalize the underlying principles for diagnosing
similar conditions in other patients.
Legal Reasoning: In legal systems, EBL can help in understanding the
principles behind legal precedents and applying them to new cases with similar
circumstances.
Automated Planning: EBL is useful in automated planning systems, where it
can learn from successful plans and generalize the steps required to achieve
similar goals in different contexts.
Natural Language Processing: EBL can be applied in natural language
processing tasks where understanding the structure and meaning behind
language is more important than statistical correlations.
Advantages of Explanation-Based Learning
Efficiency in Learning: EBL can learn effectively from a single example,
making it efficient in situations where data is scarce or expensive to obtain.
Understanding and Generalization: EBL focuses on understanding the
rationale behind examples, leading to more robust generalizations that can be
applied to a wide range of situations.
PLANNING AND MACHINE LEARNING
Interpretable Models: The rules or concepts learned through EBL are often
more interpretable than those learned through other methods, making it easier
to understand and trust the system’s decisions.
Challenges and Limitations
Dependency on Domain Knowledge: EBL relies heavily on accurate and
comprehensive domain knowledge. If the domain knowledge is incomplete or
incorrect, the system may generate flawed explanations and generalizations.
Limited to Well-Defined Problems: EBL is most effective in well-defined
problem domains where the rules and relationships are clear. It may struggle in
more complex or ambiguous domains.
Complexity of Explanation Generation: Generating explanations can be
computationally intensive, especially in domains with complex relationships and
a large number of features.