StackGAN and HDGAN

Fangda Han
1 min readJun 25, 2018

Everyday, there are interesting new ideas about how to use GANs to solve real world problems. These two papers try to solve the problem of generating high-resolution image from text.

The above figure from HDGAN gives a good illustration on the differences between the two papers (A vs. D).

Similarity

  • Both use pre-trained text embedding networks (RNN).
  • Both use GAN loss.
  • Both generate hierarchical-resolution images, and gradually add more details to the final high-resolution images.

Difference

  • StackGAN is a two-step algorithm, it first generates low-res image and then generates high-res image conditioned on the low-res image and text embedding.
  • So StackGAN needs two training steps.
  • HDGAN generate hierarchical-res images during in generator.
  • HDGAN has two discriminators, the traditional Pair Loss and the Image Loss.
  • HDGAN is end-to-end.

the local image loss (outputting a R i ×R i probability map to classify real or fake image patches).

Sign up to discover human stories that deepen your understanding of the world.

Free

Distraction-free reading. No ads.

Organize your knowledge with lists and highlights.

Tell your story. Find your audience.

Membership

Read member-only stories

Support writers you read most

Earn money for your writing

Listen to audio narrations

Read offline with the Medium app

No responses yet

Write a response