Bounding the test log-likelihood of generative models

Yoshua Bengio, Li Yao, Kyunghyun Cho

    Research output: Contribution to conferencePaperScientificpeer-review

    Abstract

    Several interesting generative learning algorithms involve a complex probability distribution over many random variables, involving intractable normalization constants or latent variable marginalization. Some of them may not have even an analytic expression for the unnormalized probability function and no tractable approximation. This makes it difficult to estimate the quality of these models, once they have been trained, or to monitor their quality (e.g. for early stopping) while training. A previously proposed method is based on constructing a non-parametric density estimator of the model’s probability function from samples generated by the model. We revisit this idea, propose a more efficient estimator, and prove that it provides a lower bound on the true test log-likelihood and an unbiased estimator as the number of generated samples goes to infinity, although one that incorporates the effect of poor mixing. We further propose a biased variant of the estimator that can be used reliably with a finite number of samples for the purpose of model comparison.

    Original languageEnglish
    Publication statusPublished - 1 Jan 2014
    MoE publication typeNot Eligible
    EventInternational Conference on Learning Representations - Banff, Canada
    Duration: 14 Apr 201416 Apr 2014
    Conference number: 2

    Conference

    ConferenceInternational Conference on Learning Representations
    Abbreviated titleICLR
    Country/TerritoryCanada
    CityBanff
    Period14/04/201416/04/2014

    Fingerprint

    Dive into the research topics of 'Bounding the test log-likelihood of generative models'. Together they form a unique fingerprint.

    Cite this