Aai Ia1 Que Ans
Aai Ia1 Que Ans
2. List the Significance of generative models and the Challenges with generative
models.
Ans: Significance of Generative Models
1. Realistic Content Generation: Create high-quality images, videos, audio, and text that
resemble real-world data.
2. Data Augmentation: Generate synthetic data to enhance model training, especially in
low-data scenarios.
3. Anomaly Detection: Identify unusual patterns in data, useful for fraud detection,
cybersecurity, and medical diagnostics.
4. Super-Resolution & Image Restoration: Improve image quality, denoise, and
reconstruct missing parts of images.
5. Personalized AI Assistants – Power chatbots, virtual assistants, and recommendation
systems with human-like responses.
Challenges with Generative Models
1. Mode Collapse: Some generative models (e.g., GANs) may produce limited variations
instead of diverse outputs.
2. High Computational Cost: Requires extensive resources, including powerful GPUs
and long training times.
3. Training Instability: Difficult to balance generator-discriminator training in GANs,
leading to inconsistent results.
4. Misinformation & Deepfakes: Can be misused to create fake videos, images, and text,
raising security and ethical concerns.
5. Overfitting to Training Data: Some models memorize training data instead of
generalizing, reducing effectiveness in real-world applications.
6. Safety & Security Risks: Malicious actors can use generative models for cyber threats,
voice spoofing, or AI-generated scams.
Where:
K = Number of Gaussian components
πi = Weight of each Gaussian component (mixing coefficient, summing to 1)
N (x∣μi, Σi) = Gaussian distribution with mean μi and covariance Σi
Each Gaussian component represents a cluster, and GMM assigns soft probabilities for each
data point to belong to different clusters.
GMMs are trained using the Expectation-Maximization (EM) Algorithm, which consists of:
1. Expectation (E-Step): Compute the probability of each data point belonging to each
Gaussian component.
2. Maximization (M-Step): Update parameters (means, covariances, and mixing
coefficients) to maximize likelihood.
3. Repeat until convergence.
Advantages of GMMs
Handles Overlapping Clusters – Unlike K-Means, GMM can model soft boundaries
and assign probabilities.
Captures Complex Data Distributions – Works well when clusters are not perfectly
spherical.
Probabilistic Assignments – Useful for applications requiring uncertainty estimation.
Challenges with GMMs
Computational Complexity – Requires more processing power than simpler
clustering algorithms.
Sensitive to Initialization – Poor initialization can lead to local optima.
Requires Choosing the Number of Components (K) – Needs methods like Bayesian
Information Criterion (BIC) to select the best K.
Where:
C= Set of all cliques in the graph
ψC(XC)= Potential function for clique C (captures dependencies)
Z = Partition function (normalization constant)
MRFs Working
1. Learning the Structure and Parameters
Maximum Likelihood Estimation (MLE): Learns the potential functions from data.
Markov Chain Monte Carlo (MCMC) Methods: Used for complex models where
direct computation is infeasible.
2. Inference in MRFs
Exact Inference: Uses methods like Variable Elimination and Junction Tree
Algorithm.
Approximate Inference: Uses algorithms like Gibbs Sampling and Belief
Propagation for large graphs.
Advantages of MRFs
Models Complex Dependencies – Suitable for problems with spatial and contextual
dependencies.
Handles Uncertainty Well – Works in noisy and uncertain environments.
No Need for Causal Assumptions – Unlike Bayesian Networks, MRFs do not require
a predefined causal structure.
Challenges of MRFs
Computationally Expensive – Exact inference is NP-hard for large graphs.
Difficult to Learn – Requires complex algorithms for parameter estimation.
Partition Function Calculation – Computing Z (normalization constant) is intractable
for large-scale models.
The generator aims to minimize the loss, while the discriminator tries to maximize its
classification accuracy.
Types of GANs: Several variants of GANs have been developed to improve training stability
and generate high-quality samples:
1. DCGAN (Deep Convolutional GAN): Uses CNNs instead of fully connected layers
for better image synthesis.
2. CGAN (Conditional GAN): Conditions the Generator on class labels for controlled
data generation.
3. WGAN (Wasserstein GAN): Improves stability by using Wasserstein distance instead
of cross-entropy loss.
4. Cycle GAN: Used for image-to-image translation (e.g., converting photos to paintings).
5. StyleGAN: Generates high-resolution, realistic images with fine-grained control over
style.
Applications of GANs
1. Image Generation & Super-Resolution: Generates realistic images (e.g., AI-
generated portraits, art, deepfakes). Enhances low-resolution images into high-
resolution.
2. Data Augmentation: Creates synthetic training data to improve AI model performance.
3. Video & Animation Generation: Used in deepfake videos, AI-powered animation,
and special effects.
4. Medical Imaging: Generates synthetic MRI/CT scans for medical research.
5. Music & Speech Synthesis: GANs generate AI-based music compositions and
synthetic voices.
3. WGAN (Wasserstein GAN) Architecture: WGAN solves training instability and mode
collapse in Vanilla GAN by introducing Wasserstein loss instead of Binary Cross-Entropy
loss. Instead of a Discriminator, WGAN uses a Critic, which assigns a score to real and fake
samples instead of classifying them as 0 or 1. The Earth Mover’s (Wasserstein) distance is
used to measure the difference between real and fake distributions, improving gradient flow
and training stability.
Structure:
Generator (G):
o Similar to DCGAN but optimized for stable training.
Critic (D) (No Sigmoid Activation):
o Replaces the Discriminator with a Critic that assigns scores instead of
probabilities.
o Uses Convolutional layers with no batch normalization.
o Outputs a real-valued score instead of a probability.
Training Process:
1. G generates fake samples, which are sent to the Critic.
2. Critic scores real and fake samples instead of classifying them.
3. Weight Clipping or Gradient Penalty stabilizes training.
4. Training continues until G generates high-quality samples.