Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Neural Nets for NLP - Latent Random Variables

Graham Neubig via YouTube

Overview

This course on Latent Random Variables in Neural Networks for NLP covers topics such as generative vs. discriminative models, variational autoencoders, handling discrete latent variables, and examples of variational autoencoders in NLP. The course aims to teach students about latent random variables, variational inference, VAE vs. AE, and the re-parameterization trick. The teaching method includes lectures, examples, and practical exercises. This course is intended for students and professionals interested in neural networks, natural language processing, and machine learning.

Syllabus

Intro
Discriminative vs. Generative Models • Discriminative model: calculate the probability of output given
Quiz: What Types of Variables? • In the an attentional sequence-to-sequence model using MLE/teacher forcing, are the following variables observed or latent? deterministic or random?
Why Latent Random Variable
What is Latent Random Variable Model
A Latent Variable Model
An Example (Goersch 2016)
Variational Inference
Practice
Variational Autoencoders
VAE vs. AE
Problem! Sampling Breaks Backprop
Solution: Re-parameterization Trick
Motivation for Latent Variables • Allows for a consistent latent space of sentences?
Difficulties in Training
KL Divergence Annealing • Basic idea: Multiply KL term by a constant starting at zero, then gradually increase to 1 • Result: model can learn to use z before getting penalized
Solution 2: Weaken the Decoder . But theoretically still problematic: it can be shown that the optimal strategy is to ignore z when it is not necessary (Chen et al. 2017)
Aggressive Inference Network Learning
Discrete Latent Variables?
Enumeration
Method 2: Sampling • Randomly sample a subset of configurations of z and optimize with respect to this subset
Method 3: Reparameterization (Maddison et al. 2017, Jang et al. 2017)
Variational Models of Language Processing (Miao et al. 2016) • Present models with random variables for document modeling and question answer pair selection
Controllable Text Generation (Hu et al. 2017)
Symbol Sequence Latent Variables (Miao and Blunsom 2016) • Encoder-decoder with a sequence of latent symbols

Taught by

Graham Neubig

Reviews

Start your review of Neural Nets for NLP - Latent Random Variables

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.