0% found this document useful (0 votes)
16 views11 pages

UNIT 1 Soft Computing

Uploaded by

Harsh Singhal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
16 views11 pages

UNIT 1 Soft Computing

Uploaded by

Harsh Singhal
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as DOCX, PDF, TXT or read online on Scribd
You are on page 1/ 11

UNIT 1

Soft Computing
Soft computing refers to a set of computational techniques that are
inspired by human-like thinking and reasoning processes. These
techniques are used to solve complex problems that are difficult to
address using traditional "hard" computing methods.
Soft computing is like a flexible and adaptable way of solving
problems, inspired by how our brains work. Instead of using strict
rules and precise calculations like traditional computing, soft
computing uses techniques that can handle uncertainty, imprecision,
and incomplete information.
Example of soft computing is fuzzy logic.

key differences between soft computing and hard computing

Aspect Soft Computing Hard Computing


Typically used for well-defined
Problem Tackles complex problems with
problems with precise
Solving uncertainty, imprecision, and partial truth.
mathematical models.
Includes fuzzy logic, neural networks,
Relies on precise algorithms and
Techniques evolutionary algorithms, and probabilistic
mathematical models.
reasoning.
Handles uncertain or incomplete data Requires clean, precise data for
Handling Data
effectively. accurate results.
Adapts and learns from data and Solutions are static and do not
Adaptability
experience, allowing for flexible solutions. adapt to changing conditions.
Decision Mimics human-like reasoning, making Relies on strict logic and precise
Making decisions based on approximate reasoning. calculations for decision making.
Commonly used in pattern recognition, Used in areas such as numerical
Real-world
optimization, control systems, and analysis, simulation, and scientific
Usage
decision-making applications. computing.
Components of Soft Computing

The main components of soft computing are:

1. Fuzzy Logic: Fuzzy logic deals with reasoning that is


approximate rather than exact. It allows for the handling of
imprecise and uncertain information by assigning degrees of
truth to statements rather than strictly true or false values.
2. Neural Networks: Neural networks are computational models
inspired by the structure and function of the human brain. They
consist of interconnected nodes (neurons) that process
information and learn from data. Neural networks excel at tasks
such as pattern recognition, classification, and prediction.
3. Evolutionary Computation: Evolutionary computation
techniques, such as genetic algorithms, evolutionary
programming, and swarm intelligence, are inspired by biological
evolution. They involve the use of population-based
optimization algorithms to find solutions to complex problems
by simulating processes like natural selection and genetic
variation.
4. Probabilistic Reasoning: Probabilistic reasoning involves
reasoning under uncertainty by using probability theory to
model and manipulate uncertain information. It allows for the
representation and manipulation of uncertain knowledge and is
commonly used in decision-making systems.
5. Machine Learning: Machine learning algorithms enable
systems to learn from data and make predictions or decisions
without being explicitly programmed. It encompasses various
techniques, including supervised learning, unsupervised
learning, and reinforcement learning, which are used to train
models on large datasets.
Structure and Function of Biological Neuron:

Structure: Biological neurons are the fundamental building blocks of


the nervous system. They consist of several key components:

1. Cell Body (Soma): The cell body contains the nucleus and other
organelles essential for the neuron's metabolic functions.
2. Dendrites: Dendrites are branching extensions that receive
signals from other neurons and transmit them towards the cell
body.
3. Axon: The axon is a long, slender projection that carries signals
away from the cell body towards other neurons, muscles, or
glands.
4. Axon Terminal: At the end of the axon, there are terminal
branches that form synaptic connections with other neurons.
These connections allow for the transmission of signals through
chemical messengers called neurotransmitters.
5. Synapse: The junction between the axon terminal of one neuron
and the dendrites or cell body of another neuron, where
neurotransmitters are released to transmit signals.

Function: Biological neurons process and transmit information in the


nervous system through electrical and chemical signaling. When a
neuron receives signals from other neurons through its dendrites,
these signals are integrated in the cell body. If the combined input
reaches a certain threshold, an action potential (a brief electrical
impulse) is generated and propagates along the axon. At the axon
terminals, neurotransmitters are released into the synapse, where they
bind to receptors on the dendrites or cell body of the next neuron, thus
transmitting the signal.

Structure and Function of (Perceptron):

Structure: Artificial neurons, also known as perceptrons, are the


basic units of artificial neural networks. They consist of:

1. Inputs: Artificial neurons receive input signals from the


environment or from other neurons.
2. Weights: Each input is multiplied by a weight, which represents
the strength of the connection between the input and the neuron.
3. Summation Function: The weighted inputs are summed
together.
4. Activation Function: The sum of the weighted inputs is passed
through an activation function, which determines the neuron's
output.
5. Output: The output of the neuron is the result of the activation
function.
Function: Artificial neurons process information in artificial neural
networks. They receive input signals, which are multiplied by
corresponding weights and summed together. The sum is then passed
through an activation function, which determines whether the neuron
should "fire" (output a signal) based on the input it received. The
output of the neuron serves as input to other neurons in the network,
propagating information through the network.

difference between biological neuron and artificial neuron

Aspect Biological Neuron Artificial Neuron (Perceptron)


Biological neurons are complex, three-
Simplified mathematical model with
dimensional structures composed of
Structure inputs, weights, summation function,
various components, including
and activation function.
dendrites, axons, and synapses
Processes information through
Information Processes information through
mathematical operations on inputs
Processing electrical and chemical signaling.
and weights.
Learns by adjusting weights based on
Learning
Learn through synaptic strength training data using algorithms like
Mechanism
backpropagation.
Limited flexibility compared to
Exhibits high flexibility and
biological neurons, primarily used for
Flexibility adaptability, enabling complex
specific tasks such as pattern
behaviors and cognitive functions.
recognition.
Artificial neurons operate at much
Biological neurons exhibit response
Response faster speeds, with response times
times on the order of milliseconds to
Time determined by computational
seconds for signaling.
processing.
Processing
Relatively slow Extremely fast, in microseconds
Speed
Size Microscopic Depends on implementation, can
range from small to large

Definition of Artificial Neural Network

An Artificial Neural Network (ANN) is a computational model


inspired by the biological neural networks found in the human brain.
It consists of interconnected nodes called artificial neurons, or simply
nodes or units. These nodes are organized in layers: an input layer,
one or more hidden layers, and an output layer.

characteristics of ANN

1. Parallel Processing: ANNs process information in a highly


parallel manner, with multiple interconnected processing units
(neurons) working simultaneously.
2. Non-linearity: ANNs are capable of modeling complex, non-linear
relationships between input and output data, making them suitable
for a wide range of tasks, including classification, regression, and
pattern recognition.
3. Learning: ANNs learn from examples through a process called
training. During training, the network adjusts its internal
parameters (weights) to minimize the difference between its
predicted outputs and the true outputs.
4. Adaptability: ANNs are highly adaptable and can learn to
recognize patterns and make predictions in a variety of domains,
from image and speech recognition to financial forecasting and
medical diagnosis.
5. Generalization: A well-trained ANN can generalize its learned
patterns to new, unseen data, making it capable of making
accurate predictions on previously unseen examples.
6. Scalability: ANNs can be scaled up to handle large datasets and
complex problems by increasing the number of neurons, layers, or
connections in the network.
7. Hardware Implementation: ANNs can be implemented in
software on traditional computing hardware or specialized
hardware accelerators like GPUs and TPUs, allowing for efficient
processing of large-scale neural networks.

Training techniques in different AN Network

1. Feedforward Neural Networks (FNNs):


Backpropagation: This is the most common training technique for
FNNs.

It involves feeding input data forward through the network to


produce an output, comparing the predicted output to the actual
output, and then adjusting the weights of the connections in the
network backward (hence "backpropagation") to minimize the
difference between the predicted and actual outputs.

2. Recurrent Neural Networks (RNNs):

Backpropagation Through Time (BPTT):

RNNs have connections that loop back on themselves, allowing


them to retain information over time.

BPTT is a variant of backpropagation specifically designed for


RNNs.

It involves unfolding the network over time (to create a


sequence of connected FNNs), applying regular
backpropagation, and then updating the weights based on the
gradients calculated at each time step.

3. Convolutional Neural Networks (CNNs):

CNNs are commonly trained using gradient descent, which


involves iteratively updating the weights of the network to
minimize a loss function (a measure of the difference between the
predicted and actual outputs).

4. Generative Adversarial Networks (GANs):

Adversarial Training: GANs consist of two networks: a generator


and a discriminator. The generator creates fake data, and the
discriminator tries to distinguish between real and fake data.
Adversarial training involves iteratively training these two
networks in opposition to each other. The generator tries to
produce more realistic data to fool the discriminator, while the
discriminator improves its ability to distinguish between real and
fake data.

5. Autoencoder Networks:

 Minimizing Reconstruction Error: Autoencoders are trained


to reconstruct their input data. Training involves minimizing the
reconstruction error, typically using techniques like mean
squared error (MSE) loss or binary cross-entropy loss.

Activation functions

Activation functions are mathematical operations applied to the


output of each neuron in an artificial neural network (ANN). They
introduce non-linearities into the network, allowing it to learn
complex patterns and relationships in the data.

ANN architectures:

1. McCulloch & Pitts Model:


 The McCulloch & Pitts model, proposed in 1943, is one of the
earliest conceptualizations of an artificial neuron. It consists of a
simple binary threshold logic unit inspired by the biological
neuron.
 The model takes binary inputs and produces a binary output based
on a threshold.
 Mathematically, it computes the weighted sum of inputs and
compares it to a threshold. If the weighted sum exceeds the
threshold, the neuron fires (outputs 1); otherwise, it remains
inactive (outputs 0).

2. Perceptron Model:
 The perceptron, introduced by Frank Rosenblatt in 1957, is an
extension of the McCulloch & Pitts model.
 It includes a linear activation n function, allowing it to process
real-valued inputs and produce real-valued outputs.
 The perceptron is trained using the perceptron learning rule, which
adjusts the weights of connections based on the error between the
predicted and actual outputs. This rule enables the perceptron to
learn to classify linearly separable patterns.

3. Linear Separability:

 Linear separability refers to the property of a dataset where


classes can be separated by a linear decision boundary.
 In the context of perceptrons, linear separability is crucial for
successful classification. If the data is linearly separable, a
perceptron can learn to classify it accurately using the
perceptron learning rule.
 If the data is not linearly separable, perceptrons may fail to
converge or produce accurate classifications.

4. ADALINE (Adaptive Linear Neuron):

 ADALINE, developed by Bernard Widrow and Ted Hoff in


1960, is an extension of the perceptron model.
 It introduces a continuous linear activation function and uses a
different learning rule called the Widrow-Hoff rule or delta rule.
 ADALINE is capable of performing linear regression tasks by
adjusting its weights to minimize the difference between the
predicted and actual outputs. It can also be used for pattern
classification tasks when combined with an appropriate decision
rule.

5. MADALINE (Multiple ADALINE):


 MADALINE, introduced by Bernard Widrow and Marcian Hoff
in 1962, is an extension of ADALINE.
 Each ADALINE unit in MADALINE is good in recognizing
different features or aspects of the input data.
 By combining the outputs of multiple ADALINE units,
MADALINE can perform more complex pattern recognition
tasks.
 MADALINE is particularly effective for tasks where the input
data is not linearly separable and requires the ability to learn
non-linear decision boundaries.

**Example: Predicting Passengers' Satisfaction**

1. **ADALINE Approach**:

- Imagine we want to predict whether passengers on a flight are


satisfied or not based on factors like flight duration and in-flight
entertainment options.

- ADALINE tries to learn a single linear relationship between these


factors and passenger satisfaction. However, if satisfaction doesn't
directly correlate with these factors in a simple linear way, ADALINE
might struggle to make accurate predictions.

2. **MADALINE Approach**:

- With MADALINE, we have multiple ADALINE units, each


focusing on different aspects of passenger satisfaction:

- One unit looks at flight duration and legroom.

- Another unit examines the quality of in-flight meals.


- A third unit considers the variety of in-flight entertainment
options.

- Each ADALINE unit learns to recognize patterns related to its


specific aspect of passenger satisfaction.

- The outputs of these units are combined to make the final


prediction about whether passengers are satisfied or not.

- For instance, if passengers tend to be satisfied when the flight is


short and they have ample legroom, regardless of entertainment
options or meal quality, the MADALINE model can capture this more
nuanced relationship and make more accurate predictions.

In this simplified example, ADALINE tries to predict satisfaction


based on simple linear relationships with less accurate predictionsss,

while MADALINE improves accuracy by considering multiple


factors simultaneously, even if these factors aren't directly related in a
linear manner.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy