Jason Liu, an applied AI consultant and creator of Instructor, discusses challenges and applications of LLMs. Topics include making LLMs interact with existing systems, building applications with LLMs, thinking in logic and design, and the future of Instructor. They also explore misconceptions in LLMs, improving LLM applications, RAG as recommendation systems, fine-tuning embedding models, measuring impact on business outcomes, and unlocking economic value through structured data extraction.
RAG framework connects with recommendation systems, utilizing language models for personalized suggestions and advanced filtering techniques beyond vector similarity.
Transfer learning and fine-tuning techniques will become more accessible and focused on task-specific models to maximize value and improve outcomes.
Businesses should prioritize evaluating impact on revenue and other tangible metrics, improving data collection and feedback for performance-driven decision-making.
Deep dives
The parallel between RAG and recommendation systems
The guest discusses the similarities between the RAG (Retrieval-Augmented Generation) framework and recommendation systems. They draw a connection to their experience at Stitch Fix, where language models were used to process user requests, filter inventory, and provide personalized clothing suggestions. They highlight how RAG utilizes a similar approach, with text chunks serving as inventory and model-generated answers acting as recommendations. They emphasize the importance of advanced filtering techniques beyond vector similarity, as well as the need for structured data extraction and feedback evaluation for improved business outcomes.
Accessible transfer learning and fine-tuning
The guest predicts that in the next year, transfer learning and fine-tuning techniques will become even more accessible and will be applied to a higher percentage of models. They believe that businesses will focus on fine-tuning models specific to their objectives and data, maximizing the value of task-specific models over generic ones. They suggest that prompt engineering will gradually give way to empirical fine-tuning practices, leveraging tailored data to achieve improved outcomes and impact.
Focused evaluation and data-driven improvement
The guest asserts that evaluation metrics and benchmarks will become less significant as businesses prioritize assessing impact on revenue, customer churn, lifetime value, etc. They highlight the need for internal evaluations that capture business-specific outcomes, rather than relying solely on general benchmarks or human evaluations. They urge companies to examine the relationships between LLM responses and tangible business metrics, as well as improve data collection and feedback to drive performance and inform decision-making.
Demystifying LLMs and focusing on design
The guest encourages demystifying LLMs and viewing them as code rather than magical systems. They argue that LLMs should be used to optimize systems and improve the assembly line of product development, rather than attempting to reinvent products or solve grand AI challenges. They emphasize the importance of good design thinking, focusing on solving concrete problems and building better user experiences, while leveraging LLM capabilities to enhance existing processes.
Challenging the status quo for improvement
The guest recommends challenging the status quo and questioning why certain products or systems are the way they are. They advocate for making small improvements and questioning existing structures, driving positive changes in one's own work and personal life. They suggest fostering a culture of continuous improvement by developing language and skills for identifying problems and implementing solutions, ultimately making a broader impact in the world.
Timestamps:
00:00 Introduction
02:18 Excitement about Machine Learning and AI
03:28 Using LLMs as Backend Developers
04:22 Building Applications with LLMs
07:07 Building Instructor
09:30 Thinking in Logic and Design
10:33 Validating Data and Building Systems with Instructor
11:49 Thoughts About Product and UX in LLMs
17:51 Future of Instructor
20:25 Misconceptions and Unsolved Problems in LLMs
24:57 Improving LLM Applications
26:14 RAG as Recommendation Systems
29:32 Fine-tuning Embedding Models
32:32 Beyond Vector Similarity in RAG
39:32 Predictions for the Next Year in AI and ML
45:26 Measuring Impact on Business Outcomes
47:06 The Continuous Cycle of Machine Learning
48:38 Unlocking Economic Value through Structured Data Extraction
50:52 Questioning the Status Quo and Making an Impact
Get the Snipd podcast app
Unlock the knowledge in podcasts with the podcast player of the future.
AI-powered podcast player
Listen to all your favourite podcasts with AI-powered features
Discover highlights
Listen to the best highlights from the podcasts you love and dive into the full episode
Save any moment
Hear something you like? Tap your headphones to save it with AI-generated key takeaways
Share & Export
Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more
AI-powered podcast player
Listen to all your favourite podcasts with AI-powered features
Discover highlights
Listen to the best highlights from the podcasts you love and dive into the full episode