METEOR metric for machine translation

This short video describes METEOR, a metric for evaluating machine generated text. It is used to evaluate whether the candidate text generated by an ML model matches the reference text (that is supposed to be generated). Where is the METEOR metric used? Meteor metric was primarily used in the Machine Translation literature. Checkout our article…

BLUE Score

This brief video describes the BLEU score, a popular evaluation metric used for sevaral tasks such as machine translation, text summarization and so on. What is BLEU Score? BLEU stands for Bilingual evaluation Understudy. It is a metric used to evaluate the quality of machine generated text by comparing it with a reference text that…

Machine Translation

Here is a high level overview of Machine Translation. This short video covers a brief history of machine translation followed by a quick explanation of SMT (statistical machine translation) and NMT (Neural Machine Translation) and a few resources to get started.

NDCG Evaluation Metric for Recommender Systems

What is the NDCG metric? NDCG stands for Normalized Discounted Cumulative Gain. Recommender systems are important in sevaral application domains such as e-commerce, finance, healthcare and so on. It is important to come up with evaluation metrics to measure how well a recommender system works. Look at our article on Evaluation metrics for recommender systems…

LIME for local explanations

This video talks about the LIME for local interpretablity. It explains the motivation for local explanations followed by how LIME works to provide local explanations along with information on where to find code examples and get started.

Global vs Local Interpretability

This video explains why we need interpretability for our AI and the two approaches typically used for interpretability namely Global and Local Interpretability. While global interpretability can give global insights about the data, it cannot explain why the prediction for a specific data point is the way it is. Local interpretability models such as LIME…

MAP at K : An evaluation metric for Ranking

This video talks about the Mean Average Precision at K (popularly called the MAP@K) metric that is commonly used for evaluating recommender systems and other ranking related problems. Why do we need Mean Average Precision@K metric? Traditional classification metrics such as precision, in the context of recommender systems can be used to see how many…

Fairness in ML: How to deal with bias in ML pipelines?

In this 30 minute video, we talk about Bias and Fairness in ML workflows: Why we need to handle fairness in ML models How biases typically creep into the ML pipeline How to measure these biases How to rectify the biases in our pipeline. A usecase with word embeddings Click here to get the latest…