Retrieval Augmented Generation (RAG) improves large language model (LLM) responses by retrieving relevant data from knowledge bases鈥攐ften private, recent, or domain-specific鈥攁nd using it to generate more accurate, grounded answers.
In this course, you鈥檒l learn how to build RAG systems that connect LLMs to external data sources. You鈥檒l explore core components like retrievers, vector databases, and language models, and apply key techniques at both the component and system level. Through hands-on work with real production tools, you鈥檒l gain the skills to design, refine, and evaluate reliable RAG pipelines鈥攁nd adapt to new methods as the field advances. Across five modules, you'll complete hands-on programming assignments that guide you through building each core part of a RAG system, from simple prototypes to production-ready components. Through hands-on labs, you鈥檒l: - Build your first RAG system by writing retrieval and prompt augmentation functions and passing structured input into an LLM. - Implement and compare retrieval methods like semantic search, BM25, and Reciprocal Rank Fusion to see how each impacts LLM responses. - Scale your RAG system using Weaviate and a real news dataset鈥攃hunking, indexing, and retrieving documents with a vector database. - Develop a domain-specific chatbot for a fictional clothing store that answers FAQs and provides product suggestions based on a custom dataset. - Improve chatbot reliability by handling real-world challenges like dynamic pricing and logging user interactions for monitoring and debugging. - Develop a domain-specific chatbot using open-source LLMs hosted by Together AI for a fictional clothing store that answers FAQs and provides product suggestions based on a custom dataset. You鈥檒l apply your skills using real-world data from domains like media, healthcare, and e-commerce. By the end of the course, you鈥檒l combine everything you鈥檝e learned to implement a fully functional, more advanced RAG system tailored to your project鈥檚 needs.