This video explores a critical flaw in reasoning Large Language Models (LLMs) called "MiP-Overthinking," where AI models produce excessively long, redundant responses when faced with ill-posed questions that have missing premises. Learn how this phenomenon affects both reinforcement learning and supervised learning models, wasting the Chain-of-Thought reasoning capabilities in advanced systems like o1, o3, and R1. Discover the research findings that reveal how current training methods fail to encourage efficient thinking, and examine the detailed analyses of reasoning length, overthinking patterns, and critical thinking locations across different LLM types. Based on research by teams from the University of Maryland, Lehigh University, KIIT Bhubaneswar, KIMS Bhubaneswar, and Monash University.
Overview
Syllabus
When Smart AI Models Overthink Stupid Data (AI TRAP)
Taught by
Discover AI