150 - How Specialized LLMs Can Help Enterprises Deliver Better GenAI User Experiences with Mark Ramsey
Aug 29, 2024
auto_awesome
Mark Ramsey from Ramsey International dives into the rapid evolution of large language models (LLMs) and their potential in enhancing user experiences. He humorously reflects on how companies can pilot LLM projects using their own website data while navigating privacy concerns. The conversation highlights the need for real human testing to improve AI interactions, the exciting prospects for healthcare, and how specialized LLMs can outperform traditional business intelligence tools. Expect insights on re-ranking strategies and the future impact of GenAI on data analytics!
The rapid evolution of generative AI necessitates organizations to remain agile and informed in leveraging the latest tools for enhanced data product management.
User experience is crucial in LLM applications, as features like streaming responses and contextual feedback significantly improve user engagement and satisfaction.
The distinction between retrieval-augmented generation and specialized models is vital for delivering accurate and relevant AI responses tailored to user needs.
Deep dives
Launch of Group Coaching Service
A new group coaching service has been launched that aims to provide accessible assistance for individuals and teams tackling UI and UX design, product management, and organizational challenges. This service is positioned as a more affordable alternative to traditional consulting, training seminars, and advisory retainers. By offering periodic outside help, the coaching is designed to help participants develop and deliver high-value data products that stand apart in the marketplace. Interested individuals can find more information through the website provided, which details the structure and benefits of the service.
Advancements in Generative AI
The rapid evolution of generative AI was highlighted, emphasizing that developments within the field feel as if weeks equate to years due to the pace of change. Recent updates, such as Google's upgraded Gemini model, represent significant improvements in the capabilities of large language models. This swift advancement underscores the necessity for organizations to stay informed and adaptable regarding the tools and solutions available for leveraging generative AI in enterprise settings. The conversation reflects a sense of urgency to integrate the latest technology effectively for enhancing data analytics and product management.
User Experience Considerations
User experience is critical when it comes to large language models, particularly in how these systems deliver feedback and responses to users. Features such as streaming responses create a sense of immediacy and engagement, improving users' perception of how fast the model is working. The discussion emphasizes the importance of providing context to users, so they feel reassured about the process until they receive an answer. Ensuring that models function seamlessly and interact in a human-like manner is vital for the acceptance and usability of generative AI in enterprise applications.
Specialized vs. Generalized Models
Two primary approaches to integrating large language models into enterprise solutions are distinguished: retrieval-augmented generation (RAG) and specialized models. RAG leverages available information around questions posed, while specialized models undergo fine-tuning to respond effectively to inquiries within specific domains, reducing the likelihood of inaccuracies or irrelevant answers. This distinction highlights the importance of tailoring AI responses to meet user needs, especially in minimizing the phenomenon of AI 'hallucinations'—where AI generates misleading information. Choosing the right approach can significantly enhance the effectiveness and trustworthiness of AI in achieving business objectives.
First Steps for Organizations
Organizations are currently in the prototype phase concerning the deployment of large language models, focusing predominantly on learning and experimentation rather than full-scale implementation. A strategic recommendation for companies is to utilize public-facing data, such as information on their websites, which can be packaged into vector databases to streamline user inquiries. This method allows organizations to develop chat interfaces that provide direct answers to customer questions, enhancing the user experience while preserving privacy. By starting with readily available information, organizations can gradually build confidence in their generative AI initiatives and demonstrate tangible value.
“Last week was a great year in GenAI,” jokes Mark Ramsey—and it’s a great philosophy to have as LLM tools especially continue to evolve at such a rapid rate. This week, you’ll get to hear my fun and insightful chat with Mark from Ramsey International about the world of large language models (LLMs) and how we make useful UXs out of them in the enterprise.
Mark shared some fascinating insights about using a company’s website information (data) as a place to pilot a LLM project, avoiding privacy landmines, and how re-ranking of models leads to better LLM response accuracy. We also talked about the importance of real human testing to ensure LLM chatbots and AI tools truly delight users. From amusing anecdotes about the spinning beach ball on macOS to envisioning a future where AI-driven chat interfaces outshine traditional BI tools, this episode is packed with forward-looking ideas and a touch of humor.
Highlights/ Skip to:
(0:50) Why is the world of GenAI evolving so fast?
(4:20) How Mark thinks about UX in an LLM application
(8:11) How Mark defines “Specialized GenAI?”
(12:42) Mark’s consulting work with GenAI / LLMs these days
(17:29) How GenAI can help the healthcare industry
(30:23) Uncovering users’ true feelings about LLM applications
(35:02) Are UIs moving backwards as models progress forward?
(40:53) How will GenAI impact data and analytics teams?
(44:51) Will LLMs be able to consistently leverage RAG and produce proper SQL?
(51:04) Where can find more from Mark and Ramsey International
Quotes from Today’s Episode
“With [GenAI], we have a solution that we’ve built to try to help organizations, and build workflows. We have a workflow that we can run and ask the same question [to a variety of GenAI models] and see how similar the answers are. Depending on the complexity of the question, you can see a lot of variability between the models… [and] we can also run the same question against the different versions of the model and see how it’s improved. Folks want a human-like experience interacting with these models.. [and] if the model can start responding in just a few seconds, that gives you much more of a conversational type of experience.” - Mark Ramsey (2:38)
“[People] don’t understand when you interact [with GenAI tools] and it brings tokens back in that streaming fashion, you’re actually seeing inside the brain of the model. Every token it produces is then displayed on the screen, and it gives you that typewriter experience back in the day. If someone has to wait, and all you’re seeing is a logo spinning, from a UX experience standpoint… people feel like the model is much faster if it just starts to produce those results in that streaming fashion. I think in a design, it’s extremely important to take advantage of that [...] as opposed to waiting to the end and delivering the results some models support that, and other models don’t.”- Mark Ramsey (4:35)
"All of the data that’s on the website is public information. We’ve done work with several organizations on quickly taking the data that’s on their website, packaging it up into a vector database, and making that be the source for questions that their customers can ask. [Organizations] publish a lot of information on their websites, but people really struggle to get to it. We’ve seen a lot of interest in vectorizing website data, making it available, and having a chat interface for the customer. The customer can ask questions, and it will take them directly to the answer, and then they can use the website as the source information.” - Mark Ramsey (14:04)
“I’m not skeptical at all. I’ve changed much of my [AI chatbot searches] to Perplexity, and I think it’s doing a pretty fantastic job overall in terms of quality. It’s returning an answer with citations, so you have a sense of where it’s sourcing the information from. I think it’s important from a user experience perspective. This is a replacement for broken search, as I really don’t want to read all the web pages and PDFs you have that *might* be about my chiropractic care query to answer my actual [healthcare] question.” - Brian O’Neill (19:22)
“We’ve all had great experience with customer service, and we’ve all had situations where the customer service was quite poor, and we’re going to have that same thing as we begin to [release more] chatbots. We need to make sure we try to alleviate having those bad experiences, and have an exit. If someone is running into a situation where they’d rather talk to a live person, have that ability to route them to someone else. That’s why the robustness of the model is extremely important in the implementation… and right now, organizations like OpenAI and Anthropic are significantly better at that [human-like] experience.” - Mark Ramsey (23:46)
"There’s two aspects of these models: the training aspect and then using the model to answer questions. I recommend to organizations to always augment their content and don’t just use the training data. You’ll still get that human-like experience that’s built into the model, but you’ll eliminate the hallucinations. If you have a model that has been set up correctly, you shouldn’t have to ask questions in a funky way to get answers.” - Mark Ramsey (39:11)
“People need to understand GenAI is not a predictive algorithm. It is not able to run predictions, it struggles with some math, so that is not the focus for these models. What’s interesting is that you can use the model as a step to get you [the answers]. A lot of the models now support functions… when you ask a question about something that is in a database, it actually uses its knowledge about the schema of the database. It can build the query, run the query to get the data back, and then once it has the data, it can reformat the data into something that is a good response back." - Mark Ramsey (42:02)