How AI Is Built  cover image

#11 Zain Hasan on Mastering Vector Databases, Product & Binary Quantization, Multi-Vector Search

How AI Is Built

NOTE

Importance of Dimension in Matryoshka Embeddings

Matryoshka embeddings involve changing the loss function to give varying importance to dimensions, unlike off-the-shelf models where all dimensions are considered equally important. Matryoshka Representation Learning assigns more importance to initial dimensions, gradually decreasing importance towards the end. This unique approach is the result of nesting loss functions from multiple models, providing different levels of information capture by each dimension. While this technique is widely used, normal binary quantization is considered more promising for its versatility. Cohere's quantization-aware embedding models show that training models to account for quantization can maintain accuracy even with quantization techniques implemented.

00:00
Transcript
Play full episode

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner