1 minute read

Paper Title Generative Adversarial Nets
Date 2014-06
Link https://arxiv.org/pdf/1703.10593.pdf

Paper summary

Reading notes

Spontaneous Questions

Q1: What’s the details of the training? A1: Several changes in the implementation such as using L2 instead of binary log loss, PatchGAN instead of GAN, etc.

Paper Review

Short Summary

The paper describes CycleGAN, an image-to-image translation approach that works very well even when there is no paired training data. In particular, CycleGAN make the translation a 2-way-street by training both generative model G and its inverse F. Beside from adversarial losses for both G and F, an additional cycle loss was employed to test the models’ ability to reconstruct the exact original image from a translation. The paper applied several techniques such as PatchGANs and L2 loss to stabilize training. It also produces quite comprehensive study of the results against the most recently developed techniques.

Strengths

  • The use of inverse function F and a cycle loss is a novel and was proven effective in producing high-quality results
  • The study of results is comprehensive thus provide credibility to the model’s claims
  • The ablation study helps provide some insights into how the model perform under different loss functions
  • Include a study with real human’s perception, which is insightful to estimate performance if applied to a real-world setting.
  • Code was released, which is helpful for re-creation

Weaknesses

  • Doesn’t have as good performance as pix2pix in many tasks, a contemporary method
  • Perception study doesn’t include the performance of pix2pix
  • The model seems to perform best on style transfer such as changing color and textual, fails to tackle more complex cases.

Reflection

  • It seems like using exact-matching for Cycle loss (L1 in this case) is not the best choice to build a robust model.
  • This model is definitely not production-ready, one thing that seems to improve a lot in later models such as diffusions

Most interesting thought/idea from reading this paper

Cycle loss is a nice idea and could be explored further. Combining with more recent techniques such as diffusions can be a good approach.

Updated: