התחל במצב לא מקוון עם האפליקציה Player FM !
פודקאסטים ששווה להאזין
בחסות


Shallow Pooling for Sparse Labels: the shortcomings of MS MARCO
Manage episode 355037191 series 3446693
In this first episode of Neural Information Retrieval Talks, Andrew Yates and Sergi Castellla discuss the paper "Shallow Pooling for Sparse Labels" by Negar Arabzadeh, Alexandra Vtyurina, Xinyi Yan and Charles L. A. Clarke from the University of Waterloo, Canada.
This paper puts the spotlight on the popular IR benchmark MS MARCO and investigates whether modern neural retrieval models retrieve documents that are even more relevant than the original top relevance annotations. The results have important implications and raise the question of to what degree this benchmark is still an informative north star to follow.
Contact: castella@zeta-alpha.com
Timestamps:
00:00 — Introduction.
01:52 — Overview and motivation of the paper.
04:00 — Origins of MS MARCO.
07:30 — Modern approaches to IR: keyword-based, dense retrieval, rerankers and learned sparse representations.
13:40 — What is "better than perfect" performance on MS MARCO?
17:15 — Results and discussion: how often are neural rankers preferred over original annotations on MS MARCO? How should we interpret these results?
26:55 — The authors' proposal to "fix" MS MARCO: shallow pooling
32:40 — How does TREC Deep Learning compare?
38:30 — How do models compare after re-annotating MS MARCO passages?
45:00 — Figure 5 audio description.
47:00 — Discussion on models' performance after re-annotations.
51:50 — Exciting directions in the space of IR benchmarking.
1:06:20 — Outro.
Related material:
- Leo Boystov paper critique blog post: http://searchivarius.org/blog/ir-leaderboards-never-tell-full-story-they-are-still-useful-and-what-can-be-done-make-them-even
- "MS MARCO Chameleons: Challenging the MS MARCO Leaderboard with Extremely Obstinate Queries" https://dl.acm.org/doi/abs/10.1145/3459637.3482011
21 פרקים
Manage episode 355037191 series 3446693
In this first episode of Neural Information Retrieval Talks, Andrew Yates and Sergi Castellla discuss the paper "Shallow Pooling for Sparse Labels" by Negar Arabzadeh, Alexandra Vtyurina, Xinyi Yan and Charles L. A. Clarke from the University of Waterloo, Canada.
This paper puts the spotlight on the popular IR benchmark MS MARCO and investigates whether modern neural retrieval models retrieve documents that are even more relevant than the original top relevance annotations. The results have important implications and raise the question of to what degree this benchmark is still an informative north star to follow.
Contact: castella@zeta-alpha.com
Timestamps:
00:00 — Introduction.
01:52 — Overview and motivation of the paper.
04:00 — Origins of MS MARCO.
07:30 — Modern approaches to IR: keyword-based, dense retrieval, rerankers and learned sparse representations.
13:40 — What is "better than perfect" performance on MS MARCO?
17:15 — Results and discussion: how often are neural rankers preferred over original annotations on MS MARCO? How should we interpret these results?
26:55 — The authors' proposal to "fix" MS MARCO: shallow pooling
32:40 — How does TREC Deep Learning compare?
38:30 — How do models compare after re-annotating MS MARCO passages?
45:00 — Figure 5 audio description.
47:00 — Discussion on models' performance after re-annotations.
51:50 — Exciting directions in the space of IR benchmarking.
1:06:20 — Outro.
Related material:
- Leo Boystov paper critique blog post: http://searchivarius.org/blog/ir-leaderboards-never-tell-full-story-they-are-still-useful-and-what-can-be-done-make-them-even
- "MS MARCO Chameleons: Challenging the MS MARCO Leaderboard with Extremely Obstinate Queries" https://dl.acm.org/doi/abs/10.1145/3459637.3482011
21 פרקים
Alle afleveringen
×
1 AGI vs ASI: The future of AI-supported decision making with Louis Rosenberg 54:42

1 EXAONE 3.0: An Expert AI for Everyone (with Hyeongu Yun) 24:57

1 Zeta-Alpha-E5-Mistral: Finetuning LLMs for Retrieval (with Arthur Câmara) 19:35

1 ColPali: Document Retrieval with Vision-Language Models only (with Manuel Faysse) 34:48

1 Using LLMs in Information Retrieval (w/ Ronak Pradeep) 22:15

1 Designing Reliable AI Systems with DSPy (w/ Omar Khattab) 59:57

1 The Power of Noise (w/ Florin Cuconasu) 11:45

1 Benchmarking IR Models (w/ Nandan Thakur) 21:55

1 Baking the Future of Information Retrieval Models 27:05

1 Hacking JIT Assembly to Build Exascale AI Infrastructure 38:04

1 The Promise of Language Models for Search: Generative Information Retrieval 1:07:31

1 Task-aware Retrieval with Instructions 1:11:13

1 Generating Training Data with Large Language Models w/ Special Guest Marzieh Fadaee 1:16:14

1 ColBERT + ColBERTv2: late interaction at a reasonable inference cost 57:30

1 Evaluating Extrapolation Performance of Dense Retrieval: How does DR compare to cross encoders when it comes to generalization? 58:30

1 Open Pre-Trained Transformer Language Models (OPT): What does it take to train GPT-3? 47:12

1 Few-Shot Conversational Dense Retrieval (ConvDR) w/ special guest Antonios Krasakis 1:23:11

1 Transformer Memory as a Differentiable Search Index: memorizing thousands of random doc ids works!? 1:01:40

1 Learning to Retrieve Passages without Supervision: finally unsupervised Neural IR? 59:10

1 The Curse of Dense Low-Dimensional Information Retrieval for Large Index Sizes 54:13

1 Shallow Pooling for Sparse Labels: the shortcomings of MS MARCO 1:07:17
ברוכים הבאים אל Player FM!
Player FM סורק את האינטרנט עבור פודקאסטים באיכות גבוהה בשבילכם כדי שתהנו מהם כרגע. זה יישום הפודקאסט הטוב ביותר והוא עובד על אנדרואיד, iPhone ואינטרנט. הירשמו לסנכרון מנויים במכשירים שונים.