DeepLearning.AI
AI is the new electricity and will transform and improve nearly all areas of human lives.

💻   Accessing Utils File and Helper Functions

In each notebook on the top menu:

1:   Click on "File"

2:   Then, click on "Open"

You will be able to see all the notebook files for the lesson, including any helper functions used in the notebook on the left sidebar. See the following image for the steps above.


💻   Downloading Notebooks

In each notebook on the top menu:

1:   Click on "File"

2:   Then, click on "Download as"

3:   Then, click on "Notebook (.ipynb)"


💻   Uploading Your Files

After following the steps shown in the previous section ("File" => "Open"), then click on "Upload" button to upload your files.


📗   See Your Progress

Once you enroll in this course—or any other short course on the DeepLearning.AI platform—and open it, you can click on 'My Learning' at the top right corner of the desktop view. There, you will be able to see all the short courses you have enrolled in and your progress in each one.

Additionally, your progress in each short course is displayed at the bottom-left corner of the learning page for each course (desktop view).


📱   Features to Use

🎞   Adjust Video Speed: Click on the gear icon (⚙) on the video and then from the Speed option, choose your desired video speed.

🗣   Captions (English and Spanish): Click on the gear icon (⚙) on the video and then from the Captions option, choose to see the captions either in English or Spanish.

🔅   Video Quality: If you do not have access to high-speed internet, click on the gear icon (⚙) on the video and then from Quality, choose the quality that works the best for your Internet speed.

🖥   Picture in Picture (PiP): This feature allows you to continue watching the video when you switch to another browser tab or window. Click on the small rectangle shape on the video to go to PiP mode.

√   Hide and Unhide Lesson Navigation Menu: If you do not have a large screen, you may click on the small hamburger icon beside the title of the course to hide the left-side navigation menu. You can then unhide it by clicking on the same icon again.


🧑   Efficient Learning Tips

The following tips can help you have an efficient learning experience with this short course and other courses.

🧑   Create a Dedicated Study Space: Establish a quiet, organized workspace free from distractions. A dedicated learning environment can significantly improve concentration and overall learning efficiency.

📅   Develop a Consistent Learning Schedule: Consistency is key to learning. Set out specific times in your day for study and make it a routine. Consistent study times help build a habit and improve information retention.

Tip: Set a recurring event and reminder in your calendar, with clear action items, to get regular notifications about your study plans and goals.

☕   Take Regular Breaks: Include short breaks in your study sessions. The Pomodoro Technique, which involves studying for 25 minutes followed by a 5-minute break, can be particularly effective.

💬   Engage with the Community: Participate in forums, discussions, and group activities. Engaging with peers can provide additional insights, create a sense of community, and make learning more enjoyable.

✍   Practice Active Learning: Don't just read or run notebooks or watch the material. Engage actively by taking notes, summarizing what you learn, teaching the concept to someone else, or applying the knowledge in your practical projects.


📚   Enroll in Other Short Courses

Keep learning by enrolling in other short courses. We add new short courses regularly. Visit DeepLearning.AI Short Courses page to see our latest courses and begin learning new topics. 👇

👉👉 🔗 DeepLearning.AI – All Short Courses [+]


🙂   Let Us Know What You Think

Your feedback helps us know what you liked and didn't like about the course. We read all your feedback and use them to improve this course and future courses. Please submit your feedback by clicking on "Course Feedback" option at the bottom of the lessons list menu (desktop view).

Also, you are more than welcome to join our community 👉👉 🔗 DeepLearning.AI Forum


Sign in

Create Your Account

Or, sign up with your email
Email Address

Already have an account? Sign in here!

By signing up, you agree to our Terms Of Use and Privacy Policy

Choose Your Learning Path

Enjoy 30% Off Now. Cancel Anytime!

MonthlyYearly

Change Your Plan

Your subscription plan will change at the end of your current billing period. You’ll continue to have access to your current plan until then.

View All Plans and Features

Welcome back!

Hi ,

We'd like to know you better so we can create more relevant courses. What do you do for work?

DeepLearning.AI
  • Explore Courses
  • Community
    • Forum
    • Events
    • Ambassadors
    • Ambassador Spotlight
  • My Learnings
  • daily streak fire

    You've achieved today's streak!

    Complete one lesson every day to keep the streak going.

    Su

    Mo

    Tu

    We

    Th

    Fr

    Sa

    free pass got

    You earned a Free Pass!

    Free Passes help protect your daily streak. Complete more lessons to earn up to 3 Free Passes.

    Free PassFree PassFree Pass
Welcome to Embedding Models: from Architecture to Implementation. Built in partnership with Vectara Embedding models create the embedding vectors that make it possible to build semantic or meaning-based retrieval systems. This course will describe their history, detailed technical architecture, and implementation. This is a technical course and so we'll focus on the building blocks rather than the applications. You may have heard of Embedding Vectors being used in Generative AI applications. These vectors have an amazing ability to capture the meaning of a word or phrase. You might have used embedding models to create these vectors, but how do these models actually work? To help us take into this, I'm delighted to introduce Amin Ahmad, Vectara's co-founder and Ofer Mendelevitch, who's the company's Head of Developer Relations. Thanks, Andrew. So glad to be here. At Vectara, we've built our own embedding models to support different RAG systems. So we have had to dig into how to select, build and train them. In this course, we will share some of the key technical details with you. That sounds great. Creating a model that can produce a vector that represents the meaning of a word is a challenging problem. You would want to make use of the large volume of existing texts as training data. But how do you go about doing that? Well, one idea was use the words around the target word as clues. Take the word tree. A text training sentence might say something like "the leaves on the tree are green", and another sentence might say "the branches on the tree are dropping". And so the words near the word tree tells you something about what tree means. If you were to have millions of sentences like that, then you might get a decent sense or get some sense of what the word tree means. This approach was made popular by an embedding model called Word2Vec by Tomáš Mikolov, Kai Chen, Greg Corrado and Jeff Dean. Most of my former teammates at Google Brain. And Word2Vec was the model that was trained on natural language to predict the word based on a few words on either side of it. Shortly after an approach called GloVe by Stanford University's Jeff Pennington, Richard Socha and Chris Manning further improved Word2Vec by simplifying the math with needed for learning embeddings. Increasing the context window around the word to produce more accurate embeddings previously involved the use of recurrent neural networks like LSTMs. The introduction of transformers in 2017 changed that, allowing feedforward neural networks, which are much more efficient to train than recurrent networks, to process sequential data effectively. The BERT model, released the following year, is deep transformer networks trained on a simple fill-in-the-missing word task to gain a deep understanding of language, ushering in the modern era of NLP and setting the stage for systems like GPT, which followed shortly after. And you can also go beyond words to longer chunks of text, like phrases or sentences. In a retrieval system, you might want to generate an embedding vector for a query sentence and compare it to vectors of response sentences. And it turns out that you can fine-tune these powerful word embedding models to evaluate sentences as well and Ofer will show you how. Exactly. In this course, you will first learn about where and how embedding models are used. Then you'll learn about BERT. This is an example of a bidirectional transformer. BERT is applied in many applications, but here we will focus on its use in retrieval. You will then learn how to build and use a contrastive loss to train a dual encoder model that is perfect for RAG applications. It has an encoder trained for queries and a separate encoder trained for responses. You will see all of this in practice. Many people have worked to create this course. I'd like to thank Vivek Sourabh from Vectara and from DeepLearning.AI, Esmaeil Gargari and Geoff Ladwig also contributed to this course. The first lesson will start off with an overview of embedding models in retrieval systems. That sounds great. Let's go on to the next video and get started.
course detail
Next Lesson
Week 1: Embedding Models: from Architecture to Implementation
  • Introduction
    Video
    ・
    4 mins
  • Introduction to embedding models
    Video
    ・
    4 mins
  • Contextualized token embeddings
    Video with Code Example
    ・
    10 mins
  • Token vs. sentence embedding
    Video with Code Example
    ・
    10 mins
  • Training a dual encoder
    Video with Code Example
    ・
    13 mins
  • Using embeddings in RAG
    Video with Code Example
    ・
    5 mins
  • Conclusion
    Video
    ・
    2 mins
  • Course Feedback
  • Community