

Benchmarking IR Models (w/ Nandan Thakur)
5 snips Jul 26, 2024
Nandan Thakur, the first author of the BEIR benchmark paper, dives deep into the realm of Information Retrieval. He discusses the current state of model evaluations, shedding light on the uphill battle of models trying to surpass the BM25 baseline. Nandan highlights the shortcomings of BEIR as of 2024 and addresses specific challenges like the Touché 2020 subset. He also shares insights on the future of benchmarking, including the newly announced TREC RAG track, and emphasizes the critical role of multilingual capabilities in advancing IR systems.
Chapters
Transcript
Episode notes