Shiva Bhattacharjee, Co-founder and CTO of TrueLaw, leverages his 20 years of tech experience to revolutionize legal workflows with bespoke models. He dives into the necessity of fine-tuning versus prompting in AI, emphasizing real-world applications in law. The discussion highlights retrieval-augmented generation and the complexities of prompt crafting for improved legal info retrieval. Additionally, he shares insights on optimizing embedding models and making strategic build vs. buy decisions in tech solutions for enhanced operational efficiency.
Read more
AI Summary
AI Chapters
Episode notes
auto_awesome
Podcast summary created with Snipd AI
Quick takeaways
Customized AI solutions in the legal sector prioritize fine-tuning over speed, enhancing relevancy and contextualization for effective information retrieval.
Establishing a flexible tech stack through existing platforms and microservices architecture simplifies workflow management while focusing on core technology optimization.
Deep dives
The Nature of AI Solutions in Law
Creating customized AI solutions for the legal sector prioritizes precision and quality over speed. The nature of legal inquiries often means that lawyers are not adept at crafting optimal prompts, making contextualization of their questions critical for effective retrieval. When implementing retrieval-augmented generation (RAG), it is essential to enhance the context of queries to improve results. By leveraging modular tools like DSPy, legal firms can adjust parameters to balance recall and precision, thus improving the relevancy of search results.
The Importance of Fine-Tuning and Data Generation
Fine-tuning embedding models tailored to specific legal content is vital for retrieving pertinent information effectively. The generation of contrast data is challenging yet essential for optimizing these models, as high-quality training data leads to improved performance. Efficiently integrating the fine-tuning process into the overall infrastructure can enhance the alignment with users’ expectations, particularly where generation approaches need to reflect legal conventions. This approach ensures that the legal solutions developed are not only relevant but also aligned with the unique presentation styles of the legal profession.
Building a Scalable and Efficient Tech Stack
The establishment of a tech stack in the AI space requires a careful assessment of whether to build or buy various components. Leveraging existing platforms for training and workflow management enables faster development cycles and reduces overhead costs associated with developing in-house solutions. Implementing a microservices architecture has proven effective, allowing asynchronous communication that supports latency-bound tasks efficiently. By utilizing proven services for workflow management, teams can focus on optimizing their core technology rather than spending time on building and maintaining infrastructure.
Shiva Bhattacharjee is the Co-founder and CTO of TrueLaw, where we are building bespoke models for law firms for a wide variety of tasks.
Alignment is Real // MLOps Podcast #260 with Shiva Bhattacharjee, CTO of TrueLaw Inc.
// Abstract
If the off-the-shelf model can understand and solve a domain-specific task well enough, either your task isn't that nuanced or you have achieved AGI. We discuss when is fine-tuning necessary over prompting and how we have created a loop of sampling - collecting feedback - fine-tuning to create models that seem to perform exceedingly well in domain-specific tasks.
// Bio
20 years of experience in distributed and data-intensive systems spanning work at Apple, Arista Networks, Databricks, and Confluent. Currently CTO at TrueLaw where we provide a framework to fold in user feedback, such as lawyer critiques of a given task, and fold them into proprietary LLM models through fine-tuning mechanics, resulting in 7-10x improvements over the base model.
// MLOps Jobs board
https://mlops.pallet.xyz/jobs
// MLOps Swag/Merch
https://mlops-community.myshopify.com/
// Related Links
Website: www.truelaw.ai
--------------- ✌️Connect With Us ✌️ -------------
Join our slack community: https://go.mlops.community/slack
Follow us on Twitter: @mlopscommunity
Sign up for the next meetup: https://go.mlops.community/register
Catch all episodes, blogs, newsletters, and more: https://mlops.community/
Connect with Demetrios on LinkedIn: https://www.linkedin.com/in/dpbrinkm/
Connect with Shiva on LinkedIn: https://www.linkedin.com/in/shivabhattacharjee/
Timestamps:
[00:00] Shiva's preferred coffee
[00:58] Takeaways
[01:17] DSPy Implementation
[04:57] Evaluating DSPy risks
[08:13] Community-driven DSPy tool
[12:19] RAG implementation strategies
[17:02] Cost-effective embedding fine-tuning
[18:51] AI infrastructure decision-making
[24:13] Prompt data flow evolution
[26:32] Buy vs build decision
[30:45] Tech stack insights
[38:20] Wrap up
Get the Snipd podcast app
Unlock the knowledge in podcasts with the podcast player of the future.
AI-powered podcast player
Listen to all your favourite podcasts with AI-powered features
Discover highlights
Listen to the best highlights from the podcasts you love and dive into the full episode
Save any moment
Hear something you like? Tap your headphones to save it with AI-generated key takeaways
Share & Export
Send highlights to Twitter, WhatsApp or export them to Notion, Readwise & more
AI-powered podcast player
Listen to all your favourite podcasts with AI-powered features
Discover highlights
Listen to the best highlights from the podcasts you love and dive into the full episode