Understand text generation: see how transformers produce output one token at a time, and why that explains so much about their behavior.
We'd like to know you better so we can create more relevant courses. What do you do for work?

Session expired â please return to Cornerstone to restart the session and complete the course.

Understand text generation: see how transformers produce output one token at a time, and why that explains so much about their behavior.
Look inside the model: build intuition for what attention is really doing, how positional encoding works, and how layers combine to make predictions.
Optimize for production: learn how quantization, KV caching, and flash attention help transformers run efficiently on GPUs.
If youâve worked with LLMs, youâve probably run into slow inference, out-of-memory errors, or hallucinations you couldnât explain. Thereâs no shortage of resources on how transformers work, but most of them either ask you to build one from scratch or get lost in theory that doesnât connect to the problems youâre actually facing.
Transformers in Practice is different. Taught by Sharon Zhou, VP of Engineering & AI at AMD, this course gives you a complete practical view of how transformers work, from how they generate text to whatâs happening inside the model to how it all gets optimized to run on real hardware. Interactive visualizations throughout let you see key concepts in action and build intuition that actually sticks.
Hereâs what youâll learn:
Youâll earn a certificate upon completing the course, recognizing your skills in transformer-based language models.
We built this course with AMD to help engineers move beyond treating LLMs as black boxes. Youâll build practical intuition for how transformers generate text, process context, and run efficiently on GPUs, while learning techniques and concepts that apply across transformer-based models and hardware environments.
This course is designed for software engineers, ML engineers, and developers who work with LLMs and want to understand whatâs actually happening under the hood.
You donât need to have built a model from scratch, but you should be comfortable using LLMs through an API or chat interface and have a basic understanding of neural network concepts like weights, layers, and training.
Gradedă»Code Assignment
Gradedă»Quiz
Gradedă»Code Assignment
Gradedă»Quiz
Gradedă»Code Assignment
Gradedă»Quiz
âWithin a few minutes and a couple slides, I had the feeling that I could learn any concept. I felt like a superhero after this course. I didnât know much about deep learning before, but I felt like I gained a strong foothold afterward.â
âThe whole specialization was like a one-stop-shop for me to decode neural networks and understand the math and logic behind every variation of it. I can say neural networks are less of a black box for a lot of us after taking the course.â
âDuring my Amazon interview, I was able to describe, in detail, how a prediction model works, how to select the data, how to train the model, and the use cases in which this model could add value to the customer.â
Please send an email to [email protected] to receive assistance.
The DeepLearning.AI Pro membership costs $25/mo billed annually and $30/mo billed monthly.
More pricing details are available on the membership page.
Important details:
Yes! If youâre a DeepLearning.AI Pro member, youâll earn a certificate upon completing the course, recognizing your skills in AI prompting.
Join today and be on the forefront of the next generation of AI!
Keep learning with updates on curated AI news, courses, events, as well as Andrewâs thoughts from DeepLearning.AI!