Evaluation Measures for Search and Recommender Systems
Offered By: James Briggs via YouTube
Course Description
Overview
Explore popular offline metrics for evaluating search and recommender systems in this 31-minute video. Learn about Recall@K, Mean Reciprocal Rank (MRR), Mean Average Precision@K (MAP@K), and Normalized Discounted Cumulative Gain (NDCG@K), with Python demonstrations for each metric. Understand the importance of evaluation measures in information retrieval systems, their impact on big tech companies' success, and how to make informed design decisions. Gain insights into dataset preparation, retrieval basics, and the pros and cons of various evaluation metrics. Access additional resources, including a related Pinecone article, code notebooks, and a discounted NLP course to further enhance your knowledge in this critical area of technology.
Syllabus
Intro
Offline Metrics
Dataset and Retrieval 101
Recall@K
Recall@K in Python
Disadvantages of Recall@K
MRR
MRR in Python
MAP@K
MAP@K in Python
NDCG@K
Pros and Cons of NDCG@K
Final Thoughts
Taught by
James Briggs
Related Courses
Artificial Intelligence for RoboticsStanford University via Udacity Intro to Computer Science
University of Virginia via Udacity Design of Computer Programs
Stanford University via Udacity Web Development
Udacity Programming Languages
University of Virginia via Udacity