Align a pretrained model for real tasks: use SFT and RLHF to improve instruction following, reasoning, and safer behavior.
We'd like to know you better so we can create more relevant courses. What do you do for work?
Instructor: Sharon Zhou
Align a pretrained model for real tasks: use SFT and RLHF to improve instruction following, reasoning, and safer behavior.
Use evaluation to guide improvements: build evals that reveal problems, choose data and rewards accordingly, and iterate.
Get models ready for production, cost-aware: plan promotion and serving, monitor reliably, and account for compute and budget.
Large language models are powerful, but raw pretrained models aren’t ready for production applications. Post-training is what adapts an LLM to follow instructions, show reasoning, and behave more safely.
Many developers still assume “LLMs inherently hallucinate,” or “only experts can tune models.” Recent advances have changed what’s feasible. If you ship LLM features (e.g., developer copilots, customer support agents, internal assistants) or work on ML/AI platform teams, understanding post-training is becoming a must-have skill.
This course, consisting of 5 modules and taught by Sharon Zhou (VP of AI at AMD and instructor to popular DeepLearning.AI courses), will guide you through various aspects of post-training:
We built this course with AMD to bring post-training practices used in leading labs to working engineers. You’ll get hands-on labs powered by AMD GPUs, while the methods you learn remain hardware-agnostic.
This course is designed for developers, ML engineers, software engineers, data scientists, and students who want to apply post-training techniques to production LLM systems. It’s also valuable for product managers and technical leaders who need to make informed decisions about post-training strategies and lead cross-functional teams working on LLM products.
To make the most of this course, we recommend strong familiarity with Python and a basic understanding of how LLMs work.
This course is part of Fine-tuning and Reinforcement Learning for LLMs: Intro to Post-Training
Graded・Quiz
Graded・Code Assignment
Graded・Code Assignment
Graded・Quiz
Graded・Code Assignment
Graded・Quiz
Graded・Code Assignment
Graded・Quiz
Graded・Code Assignment
Graded・Quiz
Graded・Code Assignment
“Within a few minutes and a couple slides, I had the feeling that I could learn any concept. I felt like a superhero after this course. I didn’t know much about deep learning before, but I felt like I gained a strong foothold afterward.”
“The whole specialization was like a one-stop-shop for me to decode neural networks and understand the math and logic behind every variation of it. I can say neural networks are less of a black box for a lot of us after taking the course.”
“During my Amazon interview, I was able to describe, in detail, how a prediction model works, how to select the data, how to train the model, and the use cases in which this model could add value to the customer.”
We recommend starting with a beginner course such as the Machine Learning Specialization.
Yes! This course is perfect for anyone with a background in Python ready to dive deeper into the post-training of large language models.
Please send an email to [email protected] to receive assistance.
A DeepLearning.AI Pro membership costs $25/month
Yes! You’ll earn a certificate upon completing the course, recognizing your skills in post-training large language models
Join today and be on the forefront of the next generation of AI!
Keep learning with updates on curated AI news, courses, events, as well as Andrew’s thoughts from DeepLearning.AI!