Unit 1 Notes
Unit 1 Notes
UNIT-I
• ML powers AI applications like natural language processing (NLP), computer vision, speech
recognition, and robotics.
a. Healthcare
• Disease diagnosis: Predict diseases using patient data (e.g., cancer detection, diabetes prediction).
• Drug discovery: Accelerate the process of finding new drugs through predictive modeling.
• Personalized medicine: Tailor treatments to individuals based on genetic data.
b. Finance
• Dynamic pricing: Adjust prices based on market trends and customer demand.
d. Manufacturing
e. Transportation
• Autonomous vehicles: Enable self-driving cars using computer vision and decision-making
algorithms.
• Traffic management: Optimize traffic flow and reduce congestion.
f. Education
g. Energy
3. Emerging Technologies
• Deep Learning: Advanced neural networks for solving complex problems in vision, NLP, and
robotics.
• Edge AI: Deploying ML models on devices like smartphones, IoT devices, and sensors.
• Quantum Machine Learning: Leveraging quantum computing to solve high-dimensional ML
problems.
• Language translation: Break language barriers with tools like Google Translate
Future Scope
• Democratization of ML: Easier access to ML tools and platforms for non-experts.
• AI in governance: Improving policymaking and public services.
• Augmented reality (AR) and virtual reality (VR): Enhancing immersive experiences.
With continuous advancements, machine learning is expected to transform industries, redefine how we work,
and address global challenges in the coming decades.
1. Pre-1950s: Foundations
• Mathematical Foundations:
• ML’s roots lie in statistics and probability theory, developed in the 18th and 19th centuries (e.g.,
Bayes’ theorem in the 1760s).
• Automata Theory:
• In the 1940s, researchers like Warren McCulloch and Walter Pitts developed a theoretical model of
artificial neurons, laying the foundation for neural networks.
• Turing proposed the idea of “learning machines” in his paper “Computing Machinery and
Intelligence” and introduced the Turing Test.
• Symbolic AI (1950s–60s):
• Early AI focused on symbolic reasoning and logic rather than learning from data (e.g., early expert
systems).
• Limitations of Perceptrons (1969):
MACHINE LEARNING NOTES
• Marvin Minsky and Seymour Papert’s book Perceptrons showed that single-layer perceptrons
couldn’t solve non-linear problems, causing a decline in neural network research.
• Bayesian Methods:
• Statistical methods, including Bayesian networks, began gaining traction for probabilistic
reasoning.
• Expert Systems:
• Rule-based systems, like MYCIN, became popular for decision-making in specific domains like
medicine.
• Evolutionary Algorithms:
• Ensemble Methods:
• Techniques like Bagging (1994) and Boosting (1997) improved predictive performance by
combining multiple models.
• Reinforcement Learning:
• Sutton and Barto popularized Reinforcement Learning, with applications in robotics and gaming.
• The internet and advancements in storage technology led to an explosion of data, creating
opportunities for ML.
• Neural networks were extended with more layers, leading to the rise of deep learning.
• Unsupervised Learning:
• Algorithms like k-means clustering and Gaussian Mixture Models became prominent for
exploring unlabeled data.
• Kernel Methods:
• Kernelized approaches allowed SVMs to handle non-linear problems efficiently.
MACHINE LEARNING NOTES
• AlexNet (2012):
• Alex Krizhevsky’s deep convolutional network won the ImageNet competition, proving deep
learning’s potential.
• Generative Models:
• Techniques like GANs (2014) and Variational Autoencoders (VAEs) emerged for generating
realistic images, videos, and text.
• DeepMind’s AlphaGo (2016) defeated the world champion in Go, showcasing the power of
combining deep learning with reinforcement learning.
• Cloud ML Platforms:
• Companies like Google, Amazon, and Microsoft launched cloud-based ML services, democratizing
access to ML tools.
• GPT models (e.g., ChatGPT) and similar architectures revolutionized conversational AI and
generative tasks.
• Self-supervised Learning:
• Models learn patterns from unlabeled data, reducing the need for expensive labeling.
• Edge AI:
• ML models are deployed on edge devices like smartphones, enabling real-time processing.
• Ethical AI:
• Multi-modal Learning:
• Models like OpenAI’s CLIP can process multiple types of data (e.g., images and text)
simultaneously.
MACHINE LEARNING NOTES
Key Milestones in Evolution:
Conclusion
Machine Learning has evolved from theoretical concepts to a transformative force impacting almost every
domain. With continued advancements, ML is poised to become even more pervasive, solving complex
global challenges and reshaping industries.
o If the given shape has four sides, and all the sides are
equal, then it will be labelled as a Square.
o If the given shape has three sides, then it will be labelled
as a triangle.
o If the given shape has six equal sides then it will be
labelled as hexagon.
Now, after training, we test our model using the test set, and the
task of the model is to identify the shape.
Advertisement
1. Regression
o Linear Regression
o Regression Trees
o Non-Linear Regression
o Bayesian Linear Regression
o Polynomial Regression
2. Classification
Spam Filtering,
o Random Forest
o Decision Trees
o Logistic Regression
MACHINE LEARNING NOTES
o Support vector Machines
o K-means clustering
o KNN (k-nearest neighbors)
o Hierarchal clustering
o Anomaly detection
o Neural Networks
o Principle Component Analysis
o Independent Component Analysis
o Apriori algorithm
MACHINE LEARNING NOTES
o Singular value decomposition
o K-means clustering
o KNN (k-nearest neighbors)
o Hierarchal clustering
o Anomaly detection
o Neural Networks
o Principle Component Analysis
MACHINE LEARNING NOTES
o Independent Component Analysis
o Apriori algorithm
o Singular value decomposition
Now, we will move further to the 6 th block, and here agent may
change the route because it always tries to find the optimal path.
So now, let's consider from the block next to the fire pit.
MACHINE LEARNING NOTES
Now, the agent has three options to move; if he moves to the blue
box, then he will feel a bump if he moves to the fire pit, then he
will get the -1 reward. But here we are taking only positive
rewards, so for this, he will move to upwards only. The complete
block values will be calculated using this formula. Consider the
below image:
MACHINE LEARNING NOTES
o Positive Reinforcement
o Negative Reinforcement
Positive Reinforcement:
This type of reinforcement can sustain the changes for a long time,
but too much positive reinforcement may lead to an overload of
states that can reduce the consequences.
Negative Reinforcement:
Markov Property:
It says that "If the agent is present in the current state S1,
performs an action a1 and move to the state s2, then the state
transition from s1 to s2 only depends on the current state and
future action and states do not depend on past actions, rewards,
or states."
Finite MDP:
A finite MDP is when there are finite states, finite rewards, and
finite actions. In RL, we consider only the finite MDP.
Markov Process:
Markov Process is a memoryless process with a sequence of random
states S1, S2, ....., St that uses the Markov Property. Markov
process is also known as Markov chain, which is a tuple (S, P) on
state S and transition function P. These two components (S and P)
can define the dynamics of the system.
o Q-Learning:
o Q-learning is an Off policy RL algorithm, which is
used for the temporal difference Learning. The
temporal difference learning methods are the way of
comparing temporally successive predictions.
MACHINE LEARNING NOTES
o It learns the value function Q (S, a), which means how
good to take action "a" at a particular state "s."
o The below flowchart explains the working of Q-
learning:
Q-Learning Explanation:
o Q-learning is a popular model-free reinforcement learning
algorithm based on the Bellman equation.
o The main objective of Q-learning is to learn the policy
which can inform the agent that what actions should be
taken for maximizing the reward under what circumstances.
o It is an off-policy RL that attempts to find the best
action to take at a current state.
o The goal of the agent in Q-learning is to maximize the
value of Q.
o The value of Q-learning can be derived from the Bellman
equation. Consider the Bellman equation given below:
In the above image, we can see there is an agent who has three
values options, V(s1), V(s2), V(s3). As this is MDP, so agent only
cares for the current state and the future state. The agent can go
to any direction (Up, Left, or Right), so he needs to decide where
to go for the optimal path. Here agent will take a move as per
probability bases and changes the state. But if we want some exact
moves, so for this, we need to make some changes in terms of Q-
value. Consider the below image:
MACHINE LEARNING NOTES
To perform any action, the agent will get a reward R(s, a), and
also he will end up on a certain state, so the Q -value equation
will be:
In Supervised learning,
RL helps to take decisions
decisions are made when input is
sequentially.
given.
MACHINE LEARNING NOTES
1. Robotics:
1. RL is used in Robot navigation, Robo-soccer, walking,
juggling, etc.
2. Control:
1. RL can be used for adaptive control such as Factory processes,
admission control in telecommunication, and Helicopter pilot
is an example of reinforcement learning.
3. Game Playing:
1. RL can be used in Game playing such as tic-tac-toe, chess,
etc.
4. Chemistry:
1. RL can be used for optimizing the chemical reactions.
5. Business:
1. RL is now used for business strategy planning.
MACHINE LEARNING NOTES
6. Manufacturing:
1. In various automobile manufacturing companies, the robots use
deep reinforcement learning to pick goods and put them in
some containers.
7. Finance Sector:
1. The RL is currently used in the finance sector for evaluating
trading strategies.
Conclusion:
From the above discussion, we can say that Reinforcement Learning
is one of the most interesting and useful parts of Machine learning.
In RL, the agent explores the environment by exploring it without
any human intervention. It is the main learning algorithm that is
used in Artificial Intelligence. But there are some cases where it
should not be used, such as if you have enough data to solve the
problem, then other ML algorithms can be used more efficiently.
The main issue with the RL algorithm is that some of the parameters
may affect the speed of the learning, such as delayed feedback.