0% found this document useful (0 votes)
25 views7 pages

AAI Extra

The document discusses challenges in generative models, including mode collapse, training instability, high computational costs, evaluation difficulties, and data bias. It also compares generative and discriminative models, highlighting their definitions, purposes, and examples. Additionally, it provides notes on CycleGAN and DCGAN, explaining their architectures, key features, and applications.

Uploaded by

Abhishek Bapat
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
0% found this document useful (0 votes)
25 views7 pages

AAI Extra

The document discusses challenges in generative models, including mode collapse, training instability, high computational costs, evaluation difficulties, and data bias. It also compares generative and discriminative models, highlighting their definitions, purposes, and examples. Additionally, it provides notes on CycleGAN and DCGAN, explaining their architectures, key features, and applications.

Uploaded by

Abhishek Bapat
Copyright
© © All Rights Reserved
We take content rights seriously. If you suspect this is your content, claim it here.
Available Formats
Download as PDF, TXT or read online on Scribd
You are on page 1/ 7

Here is a concise answer suitable for a 5-mark question on the challenges in generative

models:

✅ Challenges in Generative Models:


1.​ Mode Collapse​

○​ The model generates limited varieties of outputs, ignoring large parts of the
data distribution.​

○​ Common in GANs where the generator produces very similar outputs


regardless of input.​

2.​ Training Instability​

○​ Generative models, especially GANs, can be highly sensitive to


hyperparameters and architecture choices.​

○​ The adversarial nature can lead to non-convergence or oscillating behavior.​

3.​ High Computational Cost​

○​ Training large generative models requires significant GPU resources and


time.​

○​ Sampling from complex models (like autoregressive or diffusion models) can


be slow.​

4.​ Evaluation Difficulty​

○​ Unlike classifiers, it is hard to objectively evaluate the quality of generated


outputs.​

○​ Metrics like Inception Score or FID are approximations and may not reflect
real-world usefulness.​

5.​ Data Bias and Ethics​

○​ Generative models can amplify biases present in training data, producing


unfair or offensive content.​

○​ Managing copyright, misinformation, and deepfakes is a growing concern.​


Here’s a clear comparison table differentiating Generative and Discriminative models:

Aspect Generative Models Discriminative Models

Definition Models joint probability P(X, Y) Models conditional probability


**P(Y

Purpose Learn how data is generated Learn boundaries between


classes

Goal Generate new data points Classify or predict labels

Data Usage Understand both input X and Focus mainly on output Y


output Y given X

Examples Naive Bayes, Hidden Markov Logistic Regression, SVM,


Model, GANs, VAEs Neural Networks

Can Generate ✅ Yes ❌ No


Data?

Performance Lower classification accuracy than Higher accuracy in


(usually) discriminative classification tasks

Training Often more complex Comparatively simpler


Complexity

Overfitting Risk Higher (due to modeling full Lower


distribution)

Use Cases Image generation, speech Spam detection, sentiment


synthesis analysis
Here is a clear comparison table differentiating Virtual Reality (VR) and Augmented
Reality (AR):

Aspect Virtual Reality (VR) Augmented Reality (AR)

Definition Creates a completely Overlays digital content onto the


immersive digital environment real world

Environment Fully virtual; real world is Real world remains visible and
blocked enhanced

Device Used VR headsets (e.g., Oculus Rift, AR glasses, smartphones, tablets


HTC Vive) (e.g., HoloLens, ARKit)

User Interaction Interacts within the virtual world Interacts with real world + virtual
elements

Real World Completely replaces real Enhances the real environment


Presence surroundings

Applications Gaming, training simulations, Navigation, retail, education,


virtual tours industrial maintenance

Hardware Requires powerful hardware Can work on smartphones and


Requirement and sensors lighter devices

Level of High – isolates the user from Low to medium – blends digital
Immersion the real world with reality

Examples VR racing game, virtual surgery Pokémon Go, IKEA Place app,
simulation Snapchat filters
Q) CycleGAN

🌀 Short Note on CycleGAN


CycleGAN (Cycle-Consistent Generative Adversarial Network) is a type of generative
model designed for image-to-image translation without needing paired data.

🔑 Key Concepts:
●​ Unpaired Translation: Unlike traditional GANs that need matching image pairs (e.g.,
horse → zebra), CycleGAN learns from unpaired datasets (a collection of horse
images and a separate collection of zebra images).​

●​ Cycle Consistency: It ensures that translating an image to the target domain and
back again should give the original image. This is done using two generators:​

○​ G:X→YG: X \rightarrow Y​

○​ F:Y→XF: Y \rightarrow X​

○​ With cycle consistency loss:​


F(G(x))≈xF(G(x)) \approx x and G(F(y))≈yG(F(y)) \approx y​

●​ Discriminators: Two discriminators are used:​

○​ DXD_X for domain X (e.g., horses)​

○​ DYD_Y for domain Y (e.g., zebras)​

📷 Applications:
●​ Style transfer (e.g., turning photos into Van Gogh-style art)​

●​ Day ↔ Night conversion​

●​ Photo ↔ Painting translation​

●​ Horse ↔ Zebra transformation​


CycleGAN has two generators—G and F.
●​ G transforms images from domain X (e.g., photos) to domain Y (e.g., artwork).
●​ F transforms images from domain Y back to domain X.

Discriminators:​
There are two discriminators—Dₓ and Dᵧ.
●​ Dₓ distinguishes between real images from X and generated images from F(y).
●​ Dᵧ distinguishes between real images from Y and generated images from G(x)

Q) DCGAN

🔷 Short Note on DCGAN (Deep Convolutional GAN)


DCGAN stands for Deep Convolutional Generative Adversarial Network. It is a type of
GAN architecture that uses convolutional neural networks (CNNs) to generate realistic
images.

🔑 Key Features:
●​ Uses CNNs instead of fully connected layers to better capture spatial hierarchies in
image data.​

●​ Introduced architectural guidelines like:​

○​ No pooling layers: Use strided convolutions and transposed convolutions.​

○​ Batch normalization in both generator and discriminator for stable training.​

○​ Leaky ReLU in the discriminator and ReLU in the generator (except output
layer).​

●​ Generator learns to produce images from random noise (latent space).​

●​ Discriminator tries to distinguish between real and fake images.​


🧠 Why It’s Important:
DCGAN is foundational in generative deep learning. It improved training stability and
image quality compared to earlier GANs, and is often used as a baseline for research.

📷 Applications:
●​ Generating realistic faces​

●​ Artwork creation​

●​ Data augmentation​

●​ Style blending​

You might also like