Machine Learning Street Talk (MLST) cover image

Jay Alammar on LLMs, RAG, and AI Engineering

Machine Learning Street Talk (MLST)

NOTE

Train for Precision in Semantic Search

Recent advancements in semantic search optimization stem from the enhancement of embedding techniques since the introduction of BERT. While BERT primarily emphasizes token representations, its application for retrieval necessitates additional training, notably through contrastive training methods that distinguish between similar and dissimilar sentences. Recognizing that questions and their answers do not always align semantically necessitates further training on diverse data. This approach ensures that queries are positioned accurately within the embedding space to yield closer matches to their answers. Additionally, integrating re-rankers into the pipeline leverages the capabilities of large language models, driving further improvements in the accuracy of search outcomes.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner