Overview
Learn to create a private document AI assistant in this comprehensive tutorial that demonstrates building a local Retrieval-Augmented Generation (RAG) system. Master the implementation of advanced features including document chunking, contextual retrieval, semantic search, and integration with DeepSeek-R1 through Ollama and LangChain. Follow along to develop a system capable of analyzing PDFs and various document formats while maintaining complete data privacy. Explore key concepts including file ingestion processes, chatbot development with Ollama and LangGraph workflow, streaming capabilities, source tracking, chat history management, and Streamlit UI implementation. Get hands-on experience testing the RAG system with blog post analysis and understand the complete architecture from project structure to final deployment.
Syllabus
00:00 - Demo
00:36 - Welcome
02:01 - Architecture of our RAG
04:50 - Live "AI Engineering" Boot Camp on MLExpert.io
05:45 - Project structure and config
07:30 - Uploading files
08:21 - File ingestion retrieval - chunking, contextual retrieval, embeddings, bm25, reranking
16:49 - Chatbot Ollama, LangGraph workflow, streaming, sources, chat history
23:56 - App UI with Streamlit
26:35 - Test our RAG chat with blog post
29:14 - Conclusion
Taught by
Venelin Valkov