0% found this document useful (0 votes)
67 views17 pages

Aai Ia1 Que Ans

Generative models in AI learn the underlying distribution of datasets to create new data samples, playing a crucial role in content creation, data augmentation, and anomaly detection. Key types include Gaussian Mixture Models, Hidden Markov Models, and Bayesian Networks, each with unique advantages and challenges. Generative Adversarial Networks (GANs) consist of a Generator and Discriminator that compete to produce realistic data, with various architectures like DCGAN and WGAN enhancing performance.

Uploaded by

nelvai.lokeshwar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
67 views17 pages

Aai Ia1 Que Ans

Generative models in AI learn the underlying distribution of datasets to create new data samples, playing a crucial role in content creation, data augmentation, and anomaly detection. Key types include Gaussian Mixture Models, Hidden Markov Models, and Bayesian Networks, each with unique advantages and challenges. Generative Adversarial Networks (GANs) consist of a Generator and Discriminator that compete to produce realistic data, with various architectures like DCGAN and WGAN enhancing performance.

Uploaded by

nelvai.lokeshwar
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 17

Advanced Artificial Intelligence

Chapter 1. Generative and Probabilistic Models


1. Explain generative models and their importance in AI
Ans: Generative models in AI are a class of machine learning models that learn the underlying
distribution of a dataset and generate new data samples that resemble the original data. Unlike
discriminative models, which classify or predict labels, generative models create new content,
such as images, text, music, or even synthetic human voices. Some well-known generative
models include: Generative Adversarial Networks (GANs), Variational Autoencoders
(VAEs), Diffusion Models, Autoregressive Models (e.g., GPT, PixelRNN, WaveNet)
Importance of Generative Models in AI
1. Content Creation: Generate realistic images, videos, and audio (e.g., deepfake
technology, synthetic speech). Used in art, music, and entertainment for AI-generated
creative works.
2. Data Augmentation: Helps in scenarios where real data is limited by creating synthetic
training examples. Improves the performance of AI models in fields like medical
imaging and NLP.
3. Anomaly Detection & Fraud Detection: Generative models can learn normal patterns
in data and detect anomalies or fake data. Useful in cybersecurity, finance, and medical
diagnostics.
4. Natural Language Processing (NLP) and Conversational AI: Models like GPT-4
can generate human-like text, improving chatbots, translation, and summarization.
5. Super-Resolution and Image Inpainting: Enhance image quality by filling missing
details or converting low-resolution images into high-resolution ones.
6. Deepfake Detection and AI Ethics: Generative models can also be used to detect
deepfakes and manipulated content, ensuring authenticity in digital media.

2. List the Significance of generative models and the Challenges with generative
models.
Ans: Significance of Generative Models
1. Realistic Content Generation: Create high-quality images, videos, audio, and text that
resemble real-world data.
2. Data Augmentation: Generate synthetic data to enhance model training, especially in
low-data scenarios.
3. Anomaly Detection: Identify unusual patterns in data, useful for fraud detection,
cybersecurity, and medical diagnostics.
4. Super-Resolution & Image Restoration: Improve image quality, denoise, and
reconstruct missing parts of images.
5. Personalized AI Assistants – Power chatbots, virtual assistants, and recommendation
systems with human-like responses.
Challenges with Generative Models
1. Mode Collapse: Some generative models (e.g., GANs) may produce limited variations
instead of diverse outputs.
2. High Computational Cost: Requires extensive resources, including powerful GPUs
and long training times.
3. Training Instability: Difficult to balance generator-discriminator training in GANs,
leading to inconsistent results.
4. Misinformation & Deepfakes: Can be misused to create fake videos, images, and text,
raising security and ethical concerns.
5. Overfitting to Training Data: Some models memorize training data instead of
generalizing, reducing effectiveness in real-world applications.
6. Safety & Security Risks: Malicious actors can use generative models for cyber threats,
voice spoofing, or AI-generated scams.

3. Explain Gaussian Mixture Models (GMMs).


Ans: A Gaussian Mixture Model (GMM) is a probabilistic model that represents a dataset
as a mixture of multiple Gaussian distributions. It is commonly used for unsupervised
learning, especially in clustering, density estimation, and anomaly detection.
GMM assumes that data points are generated from multiple Gaussian distributions, each
with its own mean and variance. Unlike K-Means, which assigns each point to a single cluster,
GMM assigns a probability for each data point to belong to a cluster.
Mathematical Representation
A GMM is defined as:

Where:
 K = Number of Gaussian components
 πi = Weight of each Gaussian component (mixing coefficient, summing to 1)
 N (x∣μi, Σi) = Gaussian distribution with mean μi and covariance Σi
Each Gaussian component represents a cluster, and GMM assigns soft probabilities for each
data point to belong to different clusters.
GMMs are trained using the Expectation-Maximization (EM) Algorithm, which consists of:
1. Expectation (E-Step): Compute the probability of each data point belonging to each
Gaussian component.
2. Maximization (M-Step): Update parameters (means, covariances, and mixing
coefficients) to maximize likelihood.
3. Repeat until convergence.
Advantages of GMMs
 Handles Overlapping Clusters – Unlike K-Means, GMM can model soft boundaries
and assign probabilities.
 Captures Complex Data Distributions – Works well when clusters are not perfectly
spherical.
 Probabilistic Assignments – Useful for applications requiring uncertainty estimation.
Challenges with GMMs
 Computational Complexity – Requires more processing power than simpler
clustering algorithms.
 Sensitive to Initialization – Poor initialization can lead to local optima.
 Requires Choosing the Number of Components (K) – Needs methods like Bayesian
Information Criterion (BIC) to select the best K.

4. Explain Hidden Markov Models (HMMs).


Ans: A Hidden Markov Model (HMM) is a probabilistic model used to describe sequences
of observations where the underlying system is Markovian (i.e., the future state depends only
on the present state) but contains hidden (latent) states that are not directly observable.
HMM is widely used in speech recognition, natural language processing (NLP),
bioinformatics, finance, and robotics due to its ability to model sequential data.
HMMs Working
1. Training HMMs (Learning the Model Parameters): Baum-Welch Algorithm (a form of
Expectation-Maximization) is used to estimate model parameters from observed data.
2. Evaluating HMMs (Finding the Probability of an Observation Sequence): Forward
Algorithm is used to compute the probability of an observation sequence given an HMM.
3. Decoding HMMs (Finding the Most Likely Hidden State Sequence): Viterbi Algorithm
finds the most probable sequence of hidden states given the observations.
Advantages of HMMs
 Handles Sequential Data Efficiently – Suitable for speech, text, finance, and time-
series data.
 Probabilistic & Flexible – Provides a soft classification of states rather than hard
labels.
 Robust in Noisy Environments – Can handle uncertainty in input data.
Challenges of HMMs
 Scalability Issues – Large state spaces make HMMs computationally expensive.
 Limited Expressiveness – Cannot capture complex dependencies beyond Markovian
assumptions.
 Assumes Stationarity – Assumes fixed transition probabilities, which may not hold in
real-world applications.

5. Explain Bayesian Networks.


Ans: A Bayesian Network (BN), also known as a Bayes Net or Belief Network, is a
probabilistic graphical model that represents a set of variables and their conditional
dependencies using a Directed Acyclic Graph (DAG). It is widely used in uncertainty
modeling, decision-making, and reasoning in AI.
A Bayesian Network is defined by:
 Nodes → Represent random variables (e.g., symptoms, diseases, weather conditions).
 Edges → Represent conditional dependencies between variables (i.e., cause-effect
relationships).
 Conditional Probability Distributions (CPDs) → Each node has a probability
distribution given its parent nodes.
Mathematical Representation
The joint probability distribution (JPD) of all variables in a Bayesian Network is given by:

This factorization significantly reduces computational complexity compared to a full joint


distribution.
Bayesian Networks Working
1. Structure Learning (Building the Network)
 Expert Knowledge: Manually define dependencies based on domain knowledge.
 Data-Driven Learning: Use algorithms (e.g., Hill Climbing, Constraint-Based
Learning) to learn structure from data.
2. Parameter Learning (Estimating Probabilities)
 Maximum Likelihood Estimation (MLE) or Bayesian Estimation is used to compute
conditional probabilities.
3. Inference (Querying the Network for Probabilities)
 Exact Inference: Uses algorithms like Variable Elimination and Junction Tree
Algorithm.
 Approximate Inference: Uses methods like Monte Carlo sampling when exact
inference is infeasible.
Advantages of Bayesian Networks
 Handles Uncertainty & Missing Data – Well-suited for real-world applications with
incomplete information.
 Efficient Probabilistic Reasoning – Reduces computational complexity compared to
full joint probability tables.
 Graphical Interpretability – Easy to visualize and understand causal relationships
in data.
 Supports Decision Making – Used in expert systems and AI-driven risk assessment
models.
Challenges of Bayesian Networks
 Difficult to Learn Structure from Data – Requires large datasets or domain expertise
to define relationships.
 Scalability Issues – Inference becomes computationally expensive as the number of
variables grows.
 Assumes Causal Relationships Are Known – Incorrect assumptions can lead to poor
model performance.

6. Explain Markov Random Field (MRFs).


Ans: A Markov Random Field (MRF), also known as an undirected graphical model, is a
probabilistic model that represents dependencies between variables using an undirected
graph. Unlike Bayesian Networks (which use directed edges), MRFs use undirected edges,
meaning that the dependencies are symmetric and do not imply causality.
MRFs are widely used in computer vision, image processing, NLP, statistical physics, and
AI applications requiring spatial dependencies.
Mathematical Representation of MRFs
The probability of a set of random variables X in an MRF is defined using the Gibbs
distribution:

Where:
 C= Set of all cliques in the graph
 ψC(XC)= Potential function for clique C (captures dependencies)
 Z = Partition function (normalization constant)
MRFs Working
1. Learning the Structure and Parameters
 Maximum Likelihood Estimation (MLE): Learns the potential functions from data.
 Markov Chain Monte Carlo (MCMC) Methods: Used for complex models where
direct computation is infeasible.
2. Inference in MRFs
 Exact Inference: Uses methods like Variable Elimination and Junction Tree
Algorithm.
 Approximate Inference: Uses algorithms like Gibbs Sampling and Belief
Propagation for large graphs.
Advantages of MRFs
 Models Complex Dependencies – Suitable for problems with spatial and contextual
dependencies.
 Handles Uncertainty Well – Works in noisy and uncertain environments.
 No Need for Causal Assumptions – Unlike Bayesian Networks, MRFs do not require
a predefined causal structure.
Challenges of MRFs
 Computationally Expensive – Exact inference is NP-hard for large graphs.
 Difficult to Learn – Requires complex algorithms for parameter estimation.
 Partition Function Calculation – Computing Z (normalization constant) is intractable
for large-scale models.

Chapter 2. Generative Adversarial Network


1. Explain Generative Adversarial Networks (GANs) architecture.
Ans: A Generative Adversarial Network (GAN) is a deep learning-based generative model
consisting of two neural networks—a Generator and a Discriminator—that compete against
each other in a zero-sum game. GANs are widely used for image synthesis, data
augmentation, deepfake generation, and creative AI applications.
GAN Architecture: GANs consist of two main components:
1. Generator (G) – Creates fake data samples that resemble real data.
2. Discriminator (D) – Evaluates whether a given sample is real or fake.
1. Generator (G)
 Takes random noise (zzz) as input (usually sampled from a normal or uniform
distribution).
 Uses a deep neural network (DNN) to generate data similar to the training set.
 Outputs a synthetic sample (e.g., an image, text, or audio).
2. Discriminator (D)
 Takes an input sample (real or fake).
 Uses a deep neural network to classify the sample as real (from dataset) or fake
(from Generator).
 Outputs a probability score (1 = real, 0 = fake).
GANs Working (Training Process)
GANs are trained using an adversarial process where the Generator tries to fool the
Discriminator, and the Discriminator tries to detect fake samples.
Step-by-Step Process
1. Generator generates fake samples using random noise.
2. Discriminator evaluates real and fake samples.
3. Loss is computed:
o If Discriminator correctly identifies a fake sample, the Generator is penalized.
o If Discriminator is fooled, it is penalized, and the Generator improves.
4. Backpropagation updates both networks using gradient descent.
5. This process continues until the Generator produces highly realistic data.

The generator aims to minimize the loss, while the discriminator tries to maximize its
classification accuracy.
Types of GANs: Several variants of GANs have been developed to improve training stability
and generate high-quality samples:
1. DCGAN (Deep Convolutional GAN): Uses CNNs instead of fully connected layers
for better image synthesis.
2. CGAN (Conditional GAN): Conditions the Generator on class labels for controlled
data generation.
3. WGAN (Wasserstein GAN): Improves stability by using Wasserstein distance instead
of cross-entropy loss.
4. Cycle GAN: Used for image-to-image translation (e.g., converting photos to paintings).
5. StyleGAN: Generates high-resolution, realistic images with fine-grained control over
style.
Applications of GANs
1. Image Generation & Super-Resolution: Generates realistic images (e.g., AI-
generated portraits, art, deepfakes). Enhances low-resolution images into high-
resolution.
2. Data Augmentation: Creates synthetic training data to improve AI model performance.
3. Video & Animation Generation: Used in deepfake videos, AI-powered animation,
and special effects.
4. Medical Imaging: Generates synthetic MRI/CT scans for medical research.
5. Music & Speech Synthesis: GANs generate AI-based music compositions and
synthetic voices.

2. List the difference between DCGAN, WGAN, CGAN.


Ans:

3. Explain Architecture of Vanilla GAN, DCGAN, WGAN, Conditional GAN, Cycle


GAN.
Ans: 1. Vanilla GAN Architecture: Vanilla GAN is the simplest form of GAN, consisting of
a Generator (G) and a Discriminator (D), both implemented using fully connected layers
(MLPs). The Generator learns to produce fake data similar to real data, while the Discriminator
learns to distinguish between real and fake samples. The two networks are trained adversarially
using Binary Cross-Entropy (BCE) loss. However, Vanilla GANs suffer from unstable
training and mode collapse, making them difficult to optimize.
Structure:
 Generator (G):
o Input: Random noise (zzz).
o Passes through dense layers with ReLU activation.
o Uses Tanh activation for the final output.
 Discriminator (D):
o Input: Real or fake data sample.
o Passes through dense layers with Leaky ReLU activation.
o Uses sigmoid activation to output a probability (real or fake).
Training Process:
1. G generates fake samples, and D classifies them.
2. D updates its weights based on real vs. fake classifications.
3. G updates its weights to improve fake sample realism.
4. This adversarial process continues until G produces realistic outputs.

2. DCGAN (Deep Convolutional GAN) Architecture: DCGAN improves Vanilla GAN by


replacing fully connected layers with Convolutional Neural Networks (CNNs), making it
more suitable for image generation. The Generator uses transposed convolutions for
upsampling, while the Discriminator uses strided convolutions for downsampling. This
structure removes the need for max pooling and fully connected layers, leading to better
feature extraction and more stable training.
Structure:
 Generator (G):
o Input: Random noise (zzz).
o Uses Transpose Convolutions (ConvTranspose2D) for upsampling.
o Applies Batch Normalization & ReLU activation.
o Outputs an image using Tanh activation.
 Discriminator (D):
o Input: Real or fake image.
o Uses Convolutional layers with Leaky ReLU activation.
o Uses sigmoid activation for binary classification (real/fake).
Training Process:
1. G generates images, and D classifies them.
2. D improves its classification ability, forcing G to generate better images.
3. Both networks are updated until G produces realistic images.

3. WGAN (Wasserstein GAN) Architecture: WGAN solves training instability and mode
collapse in Vanilla GAN by introducing Wasserstein loss instead of Binary Cross-Entropy
loss. Instead of a Discriminator, WGAN uses a Critic, which assigns a score to real and fake
samples instead of classifying them as 0 or 1. The Earth Mover’s (Wasserstein) distance is
used to measure the difference between real and fake distributions, improving gradient flow
and training stability.
Structure:
 Generator (G):
o Similar to DCGAN but optimized for stable training.
 Critic (D) (No Sigmoid Activation):
o Replaces the Discriminator with a Critic that assigns scores instead of
probabilities.
o Uses Convolutional layers with no batch normalization.
o Outputs a real-valued score instead of a probability.
Training Process:
1. G generates fake samples, which are sent to the Critic.
2. Critic scores real and fake samples instead of classifying them.
3. Weight Clipping or Gradient Penalty stabilizes training.
4. Training continues until G generates high-quality samples.

4. Conditional GAN (CGAN) Architecture: CGAN extends Vanilla GAN by adding a


conditional input (label) to both the Generator and Discriminator, allowing controlled data
generation. Instead of generating data randomly, CGAN ensures the output is conditioned on
a specific label (e.g., generating images of specific objects). The Discriminator also receives
the label, ensuring that the generated output matches the expected category.
Structure:
 Generator (G):
o Input: Random noise (zzz) + condition label (yyy).
o Concatenates yyy with zzz before passing through neural network.
o Uses fully connected layers (or CNNs for images).
o Outputs a synthetic sample conditioned on yyy.
 Discriminator (D):
o Input: Real/fake sample + condition label (yyy).
o Uses concatenation of data and label before classification.
o Uses Binary Cross-Entropy loss for classification.
Training Process:
1. G receives zzz and yyy to generate labeled fake samples.
2. D evaluates real vs. fake samples based on yyy.
3. Both networks update their weights to improve generation quality.

5. CycleGAN Architecture: CycleGAN is designed for image-to-image translation,


allowing transformation between two domains (e.g., turning horses into zebras) without
paired training data. It uses two Generators and two Discriminators and introduces cycle-
consistency loss, ensuring that if an image is transformed and then translated back, it should
remain close to the original.
Structure:
 Generator 1 (G1):
o Transforms images from Domain A → Domain B (e.g., horses → zebras).
 Generator 2 (G2):
o Transforms images from Domain B → Domain A (e.g., zebras → horses).
 Discriminator 1 (D1):
o Distinguishes between real and fake images in Domain A.
 Discriminator 2 (D2):
o Distinguishes between real and fake images in Domain B.
Training Process:
1. G1 translates images from A to B, and G2 translates them back from B to A.
2. Cycle consistency loss ensures the image can be reconstructed to its original form.
3. D1 and D2 classify real vs. fake images in their respective domains.
4. Training continues until realistic domain adaptation is achieved.
4. Describe the GAN architecture in image synthesis.
Ans: The GAN architecture for image synthesis consists of two neural networks—a
Generator (G) and a Discriminator (D)—competing in a min-max game:
1. Generator (G)
o Takes random noise (latent vector) as input.
o Uses deep neural networks (DNNs) and transposed convolutions to generate
synthetic images.
o Aims to produce images that resemble real samples from the training dataset.
2. Discriminator (D)
o A convolutional neural network (CNN) that classifies images as "real" (from
the dataset) or "fake" (generated).
o Trained to maximize accuracy in distinguishing real vs. fake images.
Training Process
 The Generator improves by minimizing the adversarial loss, fooling the
Discriminator over time.
 The Discriminator is trained simultaneously to get better at distinguishing fake from
real images.
 This adversarial training refines the Generator’s outputs, leading to high-quality image
synthesis.

5. Describe the GAN applications in image synthesis and style transfer.


Ans: Image Synthesis: GANs generate new images from scratch by learning patterns from a
dataset. Some key applications include:
 Realistic Face Generation: Models like StyleGAN create high-resolution, lifelike
human faces that don’t belong to real individuals.
 Data Augmentation: GANs generate synthetic training images to improve deep
learning models, especially in medical imaging and rare-class detection.
 Super-Resolution: SRGAN enhances low-resolution images by generating fine
details, making them sharper and clearer.
 3D Object and Scene Generation: GANs synthesize realistic 3D models from 2D
images, aiding in game development and virtual reality.
Style Transfer: GAN-based style transfer allows applying artistic styles to images while
preserving content. Key applications include:
 Neural Style Transfer (NST): Models like CycleGAN apply textures and colors from
one image (e.g., a painting) to another (e.g., a photograph) while keeping structural
details intact.
 Photo-to-Painting Conversion: GANs transform photos into artistic styles resembling
Van Gogh, Monet, or Picasso.
 Domain Adaptation: CycleGAN and StarGAN convert images across different
domains, like changing a summer landscape into a winter scene or turning a horse into
a zebra.

Chapter 3. Variational Autoencoders


1. Explain VAE architecture in detail.
Ans: A Variational Autoencoder (VAE) is a generative model that learns to represent high-
dimensional data (e.g., images) in a continuous latent space, enabling controlled synthesis of
new samples. It extends the traditional Autoencoder by incorporating probabilistic modeling,
ensuring smoother and more meaningful latent representations.
VAE Architecture Components: A VAE consists of two neural networks:
(a) Encoder (Inference Network)
 Maps an input x to a latent space representation zzz.
 Instead of encoding to a fixed vector, it learns a probabilistic distribution q(z∣x),
typically modeled as a Gaussian distribution N(μ,σ2)
 Outputs:
o Mean vector (μ): Represents the center of the distribution.
o Standard deviation (σ): Represents the spread of the distribution.
(b) Latent Space Representation
 Instead of a single deterministic encoding, VAE samples z from the learned distribution:
z=μ+σ⋅ϵ where ϵ∼N(0,I) (random noise).
 This stochastic sampling ensures that similar inputs map to similar latent
representations.
(c) Decoder (Generative Network)
 Takes a sampled z and reconstructs the original input xxx.
 Models the likelihood distribution p(x∣z), often as a Gaussian N(x^,I).
 Uses deconvolutions (transpose convolutions) or fully connected layers to generate
an output.
Training Process (Loss Function): VAE is trained using a combined loss function with two
key components:
1. Reconstruction Loss (L2 Loss or Cross-Entropy)
o Measures how well the reconstructed output x^ matches the input x.
o Encourages the Decoder to generate realistic outputs.

2. KL Divergence (Regularization Loss)


o Ensures that the latent distribution q(z∣x) is close to a standard normal prior
p(z)=N(0,I).
o This prevents overfitting and enforces smooth latent space structure.

Total VAE Loss:


where β is a weight factor controlling the trade-off between reconstruction quality and latent
space regularization.
Advantages of VAE
 Generates smooth and meaningful latent representations.
 Allows controlled synthesis (e.g., interpolating between two images).
 Provides better generalization than traditional Autoencoders.
 Regularization prevents overfitting and enhances robustness.

2. List the applications of VAEs in image generation.


Ans: 1. Generating New Images: VAEs can create entirely new images by sampling from the
learned latent space. Used in generating faces, animals, landscapes, and synthetic textures.
2. Image Denoising & Inpainting: VAEs help in denoising images by learning clean
representations and reconstructing noise-free versions. Used in image inpainting, where
missing or occluded parts of an image are reconstructed.
3. Image-to-Image Translation: Conditional VAEs (CVAE) allow transformation between
different image domains. Example: Turning sketches into photorealistic images.
4. Anomaly Detection in Images: Since VAEs learn the normal distribution of data, they can
identify anomalies (e.g., defects in medical scans or industrial products). Common in medical
imaging (MRI, X-rays) and cybersecurity.
5. Handwriting & Font Generation: VAEs can generate new handwriting styles or fonts by
learning from existing samples. Example: Generating new characters for OCR datasets.
6. 3D Object and Scene Generation: VAEs are used to generate 3D models and scenes from
2D images. Helpful in virtual reality (VR), gaming, and architecture.
3. Compare Undercomplete autoencoders, Sparse autoencoders, Denoising
autoencoders.
Ans:

4. Differentiate between VAE and autoencoders.


Ans:
5. What is autoencoders explain the types of Autoencoders Undercomplete, Sparse,
Contractive, Denoising, Variational in short.
Ans: Autoencoders (AEs) are neural networks used for unsupervised learning, primarily for
dimensionality reduction, feature extraction, and data reconstruction. They consist of an
Encoder that compresses input data into a latent representation and a Decoder that
reconstructs the input from this representation. Different types of autoencoders introduce
modifications for various use cases:
a) Undercomplete Autoencoder: It uses a bottleneck architecture where the latent
space is smaller than the input. It Focuses on learning compact and useful
representations while avoiding overfitting.
Application: Dimensionality reduction (PCA alternative), feature learning.
b) Sparse Autoencoder: It uses a sparsity constraint (L1 regularization or KL
divergence) on the latent space. It Can have a larger latent space but forces only a few
neurons to be active at a time.
Application: Feature selection, anomaly detection, biological neuron modeling.
c) Contractive Autoencoder: It adds a Jacobian penalty to enforce local invariance,
ensuring small perturbations in input lead to minimal changes in latent space. It helps
in learning robust representations by reducing sensitivity to noise.
Application: Robust feature learning, representation learning.
d) Denoising Autoencoder: It introduces noise to input data and learns to reconstruct
the original input and improves robustness and prevents overfitting.
Application: Image denoising, speech enhancement, representation learning.
e) Variational Autoencoder (VAE): Instead of a fixed latent space, models it as a
probabilistic distribution (Gaussian N(μ,σ2)). It uses KL divergence to ensure
smooth and continuous latent space for generative modeling.
Application: Image generation, synthetic data creation, anomaly detection.

You might also like

pFad - Phonifier reborn

Pfad - The Proxy pFad of © 2024 Garber Painting. All rights reserved.

Note: This service is not intended for secure transactions such as banking, social media, email, or purchasing. Use at your own risk. We assume no liability whatsoever for broken pages.


Alternative Proxies:

Alternative Proxy

pFad Proxy

pFad v3 Proxy

pFad v4 Proxy