Evaluating LLM Performance: DeepSeek, Phi-3.5, and LLaMA 3 Using Chain of Thought Reasoning
The Machine Learning Engineer via YouTube
Overview
Explore a detailed comparison of logical and mathematical reasoning capabilities across three language models - DeepSeek R1 1.5B, Microsoft Phi3.5 3.8B, and LLama 3.2 3B - in this 26-minute technical evaluation video. Learn how to implement Chain of Thought (CoT) prompting techniques using Ollama and Langchain frameworks, with all models quantized to 4Int for performance analysis. Access the complete evaluation process through the provided Jupyter notebook to understand the comparative strengths and limitations of these models in handling complex reasoning tasks.
Syllabus
LLM,s: CoT Evaluation DeepSeek R1 1.B, Phi3.5 3.8B and LLama 3.2 3B #datascience #machinelearning
Taught by
The Machine Learning Engineer