ebook img

Generative Adversarial Networks for Discrete Data PDF

65 Pages·2017·6.75 MB·English
by  
Save to my drive
Quick download
Download
Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.

Preview Generative Adversarial Networks for Discrete Data

Generative Adversarial Networks (GANs) for Discrete Data Lantao Yu Shanghai Jiao Tong University http://lantaoyu.com July 26, 2017 Self Introduction – Lantao Yu • Position • Research Assistant at CS Dept. of SJTU 2015-now • Apex Data and Knowledge Management Lab • Research on deep learning, reinforcement learning and multi-agent systems • Education • Third-year Undergraduate Student, Computer Science Dept., Shanghai Jiao Tong University, China, 2014-now • Contact • [email protected][email protected] Apex Data & Knowledge Management Lab • Machine learning and data science • with applications of recommender systems, computational ads, social networks, crowdsourcing, urban computing etc. • Students • 8 PhDs • 8 Masters • 24 Undergrads • www.apexlab.org Professor Yong Yu A.P. Weinan Zhang Content • Fundamentals - Generative Adversarial Networks • Connection and difference between generating discrete data and continuous data with GANs • Advances - GANs for Discrete Data • SeqGAN: Sequence Generation via GANs with Policy Gradient • IRGAN: A Minimax Game for Unifying Generative and Discriminative Information Retrieval Models Generative Adversarial Networks (GANs) [Goodfellow, I., et al. 2014. Generative adversarial nets. In NIPS 2014.] Problem Definition • Given a dataset D = x , build a model q ( x ) of { } p(x) the data distribution that fits the true one • Traditional objective: maximum likelihood estimation (MLE) 1 max log q(x) max [log q(x)] E x p(x) D ⇡ q q ⇠ | | x D X 2 • Check whether a true data is with a high mass density of the learned model Problems of MLE • Inconsistency of Evaluation and Use max [log q(x)] max [log p(x)] E E x p(x) x q(x) q ⇠ q ⇠ Training/evaluation Use (Turing Test) • Check whether a • Check whether a true data is with a model-generated high mass density data is considered as of the learned true as possible model • More straightforward • Approximated by but it is hard or 1 impossible to directly max log q(x) calculate p(x) D q | | x D X 2 Problems of MLE • Equivalent to minimizing asymmetric KL(p(x) q(x)) : || p(x) KL(p(x) q(x)) = p(x) log dx || q(x) Z X • When p(x) > 0 but q(x) 0, the integrand inside the KL ! grows quickly to infinity, which means MLE assigns an extremely high cost to such situation • When p(x) 0 but q(x) > 0, the value inside the KL ! divergence goes to 0, which means MLE pays extremely low cost for generating fake looking samples [ArjovskyM, BottouL. Towards principled methods for training generative adversarial networks. arXiv, 2017.] Problems of MLE • Exposure bias in sequential data generation: Training Inference Update the model as follows: When generating the next y token , sample from: t ˆ max E log G (y Y ) G (yˆ Y ) Y p ✓ t 1:t 1 ✓ t 1:t 1 ✓ ⇠ true | � | � t X The real prefix The guessed prefix Generative Adversarial Nets (GANs) • What we really want max [log p(x)] E x q(x) q ⇠ • But we cannot directly calculate pp((xx)) • Idea: what if we build a discriminator to judge whether a data instance is true or fake (artificially generated)? • Leverage the strong power of deep learning based discriminative models

Description:
of honesty and integrity and responsibility to which so many of you have committed your careers. It's a standard that's a little bit harder to find today.
See more

The list of books you might like

Most books are stored in the elastic cloud where traffic is expensive. For this reason, we have a limit on daily download.