Class Central is learner-supported. When you buy through links on our site, we may earn an affiliate commission.

YouTube

Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach

AutoML Seminars via YouTube

Overview

Coursera Plus Monthly Sale: All Certificates & Courses 40% Off!
This seminar presents a novel language model architecture that scales test-time computation through latent space reasoning. Join speaker Jonas Geiping as he explores a recurrent block approach that can unroll to arbitrary depth at test-time, contrasting with mainstream reasoning models that scale by producing more tokens. Learn about this innovative method that doesn't require specialized training data, works with small context windows, and captures reasoning types not easily represented in words. Discover how a 3.5 billion parameter model trained on 800 billion tokens can dramatically improve performance on reasoning benchmarks, achieving computation equivalent to 50 billion parameters. The presentation covers the complete research findings detailed in the associated paper available on arXiv.

Syllabus

Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach

Taught by

AutoML Seminars

Reviews

Start your review of Scaling up Test-Time Compute with Latent Reasoning: A Recurrent Depth Approach

Never Stop Learning.

Get personalized course recommendations, track subjects and courses with reminders, and more.

Someone learning on their laptop while sitting on the floor.