DeepLearning.AI
AI is the new electricity and will transform and improve nearly all areas of human lives.

๐Ÿ’ป ย  Accessing Utils File and Helper Functions

In each notebook on the top menu:

1: ย  Click on "File"

2: ย  Then, click on "Open"

You will be able to see all the notebook files for the lesson, including any helper functions used in the notebook on the left sidebar. See the following image for the steps above.


๐Ÿ’ป ย  Downloading Notebooks

In each notebook on the top menu:

1: ย  Click on "File"

2: ย  Then, click on "Download as"

3: ย  Then, click on "Notebook (.ipynb)"


๐Ÿ’ป ย  Uploading Your Files

After following the steps shown in the previous section ("File" => "Open"), then click on "Upload" button to upload your files.


๐Ÿ“— ย  See Your Progress

Once you enroll in this courseโ€”or any other short course on the DeepLearning.AI platformโ€”and open it, you can click on 'My Learning' at the top right corner of the desktop view. There, you will be able to see all the short courses you have enrolled in and your progress in each one.

Additionally, your progress in each short course is displayed at the bottom-left corner of the learning page for each course (desktop view).


๐Ÿ“ฑ ย  Features to Use

๐ŸŽž ย  Adjust Video Speed: Click on the gear icon (โš™) on the video and then from the Speed option, choose your desired video speed.

๐Ÿ—ฃ ย  Captions (English and Spanish): Click on the gear icon (โš™) on the video and then from the Captions option, choose to see the captions either in English or Spanish.

๐Ÿ”… ย  Video Quality: If you do not have access to high-speed internet, click on the gear icon (โš™) on the video and then from Quality, choose the quality that works the best for your Internet speed.

๐Ÿ–ฅ ย  Picture in Picture (PiP): This feature allows you to continue watching the video when you switch to another browser tab or window. Click on the small rectangle shape on the video to go to PiP mode.

โˆš ย  Hide and Unhide Lesson Navigation Menu: If you do not have a large screen, you may click on the small hamburger icon beside the title of the course to hide the left-side navigation menu. You can then unhide it by clicking on the same icon again.


๐Ÿง‘ ย  Efficient Learning Tips

The following tips can help you have an efficient learning experience with this short course and other courses.

๐Ÿง‘ ย  Create a Dedicated Study Space: Establish a quiet, organized workspace free from distractions. A dedicated learning environment can significantly improve concentration and overall learning efficiency.

๐Ÿ“… ย  Develop a Consistent Learning Schedule: Consistency is key to learning. Set out specific times in your day for study and make it a routine. Consistent study times help build a habit and improve information retention.

Tip: Set a recurring event and reminder in your calendar, with clear action items, to get regular notifications about your study plans and goals.

โ˜• ย  Take Regular Breaks: Include short breaks in your study sessions. The Pomodoro Technique, which involves studying for 25 minutes followed by a 5-minute break, can be particularly effective.

๐Ÿ’ฌ ย  Engage with the Community: Participate in forums, discussions, and group activities. Engaging with peers can provide additional insights, create a sense of community, and make learning more enjoyable.

โœ ย  Practice Active Learning: Don't just read or run notebooks or watch the material. Engage actively by taking notes, summarizing what you learn, teaching the concept to someone else, or applying the knowledge in your practical projects.


๐Ÿ“š ย  Enroll in Other Short Courses

Keep learning by enrolling in other short courses. We add new short courses regularly. Visit DeepLearning.AI Short Courses page to see our latest courses and begin learning new topics. ๐Ÿ‘‡

๐Ÿ‘‰๐Ÿ‘‰ ๐Ÿ”— DeepLearning.AI โ€“ All Short Courses [+]


๐Ÿ™‚ ย  Let Us Know What You Think

Your feedback helps us know what you liked and didn't like about the course. We read all your feedback and use them to improve this course and future courses. Please submit your feedback by clicking on "Course Feedback" option at the bottom of the lessons list menu (desktop view).

Also, you are more than welcome to join our community ๐Ÿ‘‰๐Ÿ‘‰ ๐Ÿ”— DeepLearning.AI Forum


Sign in

Create Your Account

Or, sign up with your email
Email Address

Already have an account? Sign in here!

By signing up, you agree to our Terms Of Use and Privacy Policy

Choose Your Learning Path

Enjoy 30% Off Now. Cancel Anytime!

MonthlyYearly

Change Your Plan

Your subscription plan will change at the end of your current billing period. Youโ€™ll continue to have access to your current plan until then.

View All Plans and Features

Welcome back!

Hi ,

We'd like to know you better so we can create more relevant courses. What do you do for work?

DeepLearning.AI
  • Explore Courses
  • Community
    • Forum
    • Events
    • Ambassadors
    • Ambassador Spotlight
  • My Learnings
  • daily streak fire

    You've achieved today's streak!

    Complete one lesson every day to keep the streak going.

    Su

    Mo

    Tu

    We

    Th

    Fr

    Sa

    free pass got

    You earned a Free Pass!

    Free Passes help protect your daily streak. Complete more lessons to earn up to 3 Free Passes.

    Free PassFree PassFree Pass
Welcome to Introducing Multimodal Llama 3.2, built in partnership with Meta and taught by Amit Sangani, who's the Director of AI Partner engineering for the Llama team at Meta. It's my pleasure to be here with you, Andrew, and to teach this course. Many of us as developers and researchers, appreciate the value of open software and models. Open models are key building blocks of AI and are key enabler of AI research. Open models, that is, models that anyone can download, customize, fine tuned, or build new applications on top of, are really important components of how AI innovation takes place. So for example, if you look on HuggingFace, you find thousands of variants of the Llama models and they represent research or different applications that others have built as a result of Meta providing open models that others could make these innovations on top of. Recently Meta launched Llama 3.2, adding new models and capabilities to the Llama family, which opens up new capabilities for developers. That's right. At Meta, we believe that openness drives innovation and is the right path forward. That is why we continue to share our research and collaborate with our partners and the developer community. In the 3.2 release, we released four new models. We added vision capability to two of the previous 3.1 models, resulting in the 3.2 11 B and 90 B models. To support edge applications, we added the smaller 1B and 3B models. This is on top of the recent 3.1 release, which included 405 billion parameter model, which is a foundation class model. One important development is that the open Llama family now has vision capabilities as well, and Llama Stack additionally provides an open source set of software to help developers built on top of the Llama models. In this course, we'll start with an introduction of the Llama family of models, how they were built and trained, and how you can use them in your applications. In addition, you might have heard of an LLMs user and assistant rules. Llama 3.2 introduces new roles, for example, the iPython rule. This is a tool calling rule and gives the Llama family built-in and user-defined function calling capabilities, which are useful for building agentic workflows. You'll get practice in the labs with the prompt format and tokenization to support this as well as many examples of tool calling. The 3.1 and 3.2 models support an expanded vocabulary of 128,000 tokens and use tiktoken tokenizer. You will also look more closely at tokenization as well as how it impacts performance. And as Andrew mentioned, you will build with the Llama Stack API. This is a standardized interface for canonical toolchain components like fine-tuning or synthetic data generation to customize Llama models and build agentic applications. Many people have worked to create this course. I'd like to thank Jeff Tang and Xi Yan from Meta. As well as Esmaeil Gargari and Geoff Ladwig from DeepLearning.AI. All right, let's set up the llamas and get started!
course detail
Next Lesson
Introducing Multimodal Llama 3.2
  • Introduction
    Video
    ใƒป
    3 mins
  • Overview of Llama 3.2
    Video
    ใƒป
    5 mins
  • Multimodal Prompting
    Video with Code Example
    ใƒป
    10 mins
  • Multimodal Use Cases
    Video with Code Example
    ใƒป
    14 mins
  • Prompt Format
    Video with Code Example
    ใƒป
    12 mins
  • Tokenization
    Video with Code Example
    ใƒป
    7 mins
  • Tool Calling
    Video with Code Example
    ใƒป
    18 mins
  • Llama Stack
    Video with Code Example
    ใƒป
    6 mins
  • Conclusion
    Video
    ใƒป
    1 min
  • Course Feedback
  • Community