Abstract

Fréchet Inception Distance (FID) is the primary metric for ranking models in data-driven generative modeling. While remarkably successful, the metric is known to sometimes disagree with human judgement. We investigate a root cause of these discrepancies, and visualize what FID "looks at" in generated images. We show that the feature space that FID is (typically) computed in is so close to the ImageNet classifications that aligning the histograms of Top-N classifications between sets of generated and real images can reduce FID substantially -- without actually improving the quality of results. Thus, we conclude that FID is prone to intentional or accidental distortions. As a practical example of an accidental distortion, we discuss a case where an ImageNet pre-trained FastGAN achieves a FID comparable to StyleGAN2, while being worse in terms of human evaluation.
Original languageEnglish
Title of host publication11th International Conference on Learning Representations (ICLR 2023)
PublisherCurran Associates Inc.
Number of pages26
ISBN (Print)9781713899259
Publication statusPublished - 1 May 2023
MoE publication typeA4 Conference publication
EventInternational Conference on Learning Representations - Kigali, Rwanda
Duration: 1 May 20235 May 2023
Conference number: 11
https://iclr.cc/

Conference

ConferenceInternational Conference on Learning Representations
Abbreviated titleICLR
Country/TerritoryRwanda
CityKigali
Period01/05/202305/05/2023
Internet address

Fingerprint

Dive into the research topics of 'The Role of ImageNet Classes in Fréchet Inception Distance'. Together they form a unique fingerprint.
  • PIPE: ERC PIPE/Lehtinen

    01/05/202031/08/2025

    Project: EU_H2ERC

Cite this