How AI Is Built  cover image

How AI Is Built

Latest episodes

undefined
Feb 6, 2025 • 1h 34min

#041 Context Engineering, How Knowledge Graphs Help LLMs Reason

Robert Caulk, who leads Emergent Methods and has over 1,000 academic citations, dives into the fascinating world of knowledge graphs and their integration with large language models (LLMs). He discusses how these graphs help AI systems connect complex data relationships, enhancing reasoning accuracy. The conversation also touches on the challenges of multilingual entity extraction and the need for context engineering to improve AI-generated content. Additionally, Caulk shares insights into upcoming features for real-time event tracking and the future of project management tools.
undefined
5 snips
Jan 31, 2025 • 52min

#040 Vector Database Quantization, Product, Binary, and Scalar

Zain Hasan, a former ML engineer at Weaviate and now a Senior AI/ML Engineer at Together, dives into the fascinating world of vector database quantization. He explains how quantization can drastically reduce storage costs, likening it to image compression. Zain discusses three quantization methods: binary, product, and scalar, each with unique trade-offs in precision and efficiency. He also addresses the speed and memory usage challenges of managing vector data, and hints at exciting future applications, including brain-computer interfaces.
undefined
Jan 23, 2025 • 53min

#039 Local-First Search, How to Push Search To End-Devices

Alex Garcia, a developer passionate about making vector search practical, discusses his creation, SQLiteVec. He emphasizes its lightweight design and how it simplifies local AI applications. The conversation reveals the efficiency of SQLiteVec's brute force searches, with impressive performance metrics at scale. Garcia also dives into challenges like data synchronization and fine-tuning embedding models. His insights on binary quantization and future innovations in local search highlight the evolution of user-friendly machine learning tools.
undefined
21 snips
Jan 9, 2025 • 1h 14min

#038 AI-Powered Search, Context Is King, But Your RAG System Ignores Two-Thirds of It

Trey Grainger, author of 'AI-Powered Search' and an expert in search systems, joins the conversation to unravel the complexities of retrieval and generation in AI. He presents the concept of 'GARRAG,' where retrieval and generation enhance each other. Trey dives into the importance of user context, discussing how behavior signals improve search personalization. He shares insights on moving from simple vector similarity to advanced models and offers practical advice for engineers on choosing effective tools, promoting a structured, modular approach for better search results.
undefined
18 snips
Jan 3, 2025 • 49min

#037 Chunking for RAG: Stop Breaking Your Documents Into Meaningless Pieces

Brandon Smith, a research engineer at Chroma known for his extensive work on chunking techniques for retrieval-augmented generation systems, shares his insights on optimizing semantic search. He discusses the common misconceptions surrounding chunk sizes and overlap, highlighting the challenges of maintaining context in dense content. Smith criticizes existing strategies, such as OpenAI's 800-token chunks, and emphasizes the importance of coherent parsing. He also introduces innovative approaches to enhance contextual integrity in document processing, paving the way for improved information retrieval.
undefined
14 snips
Dec 19, 2024 • 48min

#036 How AI Can Start Teaching Itself - Synthetic Data Deep Dive

Adrien Morisot, an ML engineer at Cohere, discusses the transformative use of synthetic data in AI training. He explores the prevalent practice of using synthetic data in large language models, emphasizing model distillation techniques. Morisot shares his early challenges in generative models, breakthroughs driven by customer needs, and the importance of diverse output data. He also highlights the critical role of rigorous validation in preventing feedback loops and the potential for synthetic data to enhance specialized AI applications across various fields.
undefined
5 snips
Dec 13, 2024 • 46min

#035 A Search System That Learns As You Use It (Agentic RAG)

Stephen Batifol, an expert in Agentic RAG and advanced search technology, dives into the future of search systems. He discusses how modern retrieval-augmented generation (RAG) systems smartly match queries to the most suitable tools, utilizing a mix of methods. Batifol emphasizes the importance of metadata and modular design in creating effective search workflows. The conversation touches on adaptive AI capabilities for query refinement and the significance of user feedback in improving system accuracy. He also addresses the challenges of ambiguity in user queries, highlighting the need for innovative filtering techniques.
undefined
Dec 5, 2024 • 47min

#034 Rethinking Search Inside Postgres, From Lexemes to BM25

Philippe Noël, Founder and CEO of ParadeDB, dives into the revolutionary shift in search technology with his open-source PostgreSQL extension. He discusses how ParadeDB eliminates the need for separate search clusters by enabling search directly within databases, simplifying architecture and enhancing cost-efficiency. The conversation explores BM25 indexing, maintaining data normalization, and the advantages of ACID compliance with search. Philippe also reveals successful use cases, including Alibaba Cloud’s implementation, and practical insights for optimizing large-scale search applications.
undefined
14 snips
Nov 28, 2024 • 51min

#033 RAG's Biggest Problems & How to Fix It (ft. Synthetic Data)

Saahil Ognawala, Head of Product at Jina AI and expert in RAG systems, dives deep into the complexities of retrieval augmented generation. He reveals why RAG systems often falter in production and how strategic testing and synthetic data can enhance performance. The conversation covers the vital role of user intent, evaluation metrics, and the balancing act between real and synthetic data. Saahil also emphasizes the importance of continuous user feedback and the need for robust evaluation frameworks to fine-tune AI models effectively.
undefined
Nov 21, 2024 • 47min

#032 Improving Documentation Quality for RAG Systems

Max Buckley, a Google expert in LLM experimentation, dives into the hidden dangers of poor documentation in RAG systems. He explains how even one ambiguous sentence can skew an entire knowledge base. Max emphasizes the challenge of identifying such "documentation poisons" and discusses the importance of multiple feedback loops for quality control. He highlights unique linguistic ecosystems in large organizations and shares insights on enhancing documentation clarity and consistency to improve AI outputs.

Remember Everything You Learn from Podcasts

Save insights instantly, chat with episodes, and build lasting knowledge - all powered by AI.
App store bannerPlay store banner
Get the app