Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

CMU Neural Nets for NLP - Models with Latent Random Variables

Graham Neubig via YouTube

Overview

Limited-Time Offer: Up to 75% Off Coursera Plus!
7000+ certificate courses from Google, Microsoft, IBM, and many more.
This course focuses on teaching neural network models for natural language processing with latent random variables. The learning outcomes include understanding the differences between discriminative and generative models, implementing the re-parameterization trick, and utilizing Gumbel-Softmax for continuous gradients. The course teaches skills such as creating optimizable objectives using KL divergence, generating language models, and training variational models for language processing. The teaching method involves lectures and examples to explain concepts and techniques. The intended audience for this course is individuals interested in neural networks, natural language processing, and incorporating latent variables into their models.

Syllabus

Intro
Discriminative vs. Generative Models • Discriminative model: calculate the probability of output given
Quiz: What Types of Variables?
Why Latent Random Variables?
An Example (Goersch 2016)
Problem: Straightforward Sampling is Inefficient
Solution: "Inference Model" • Predict which latent point produced the data point using inference
Disconnect Between Samples and Objective
VAE Objective • We can create an optimizable objective matching our problem, starting with KL divergence
Interpreting the VAE Objective
Problem! Sampling Breaks Backprop
Solution: Re-parameterization Trick
Generating from Language Models
Motivation for Latent Variables . Allows for a consistent latent space of sentences?
Difficulties in Training
KL Divergence Annealing
Weaken the Decoder
Discrete Latent Variables?
Method 1: Enumeration
Reparameterization (Maddison et al. 2017, Jang et al. 2017)
Gumbel-Softmax • A way to soften the decision and allow for continuous gradients
Variational Models of Language Processing (Miao et al. 2016)
Controllable Text Generation (Hu et al. 2017)
Symbol Sequence Latent Variables (Miao and Blunsom 2016)

Taught by

Graham Neubig

Reviews

Start your review of CMU Neural Nets for NLP - Models with Latent Random Variables

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.