Latent Space: The AI Engineer Podcast

How to train a Million Context LLM — with Mark Huang of Gradient.ai

50 snips
May 30, 2024
Mark Huang, Co-founder of Gradient.ai, dives into the exciting advancements in AI, particularly long context learning. He discusses the evolution of context lengths, mapping out a timeline of breakthroughs and innovations in LLMs. Mark reflects on his team's work with Llama 3 and the challenges of training models with vast token capacities. He also sheds light on optimizing GPU performance and the pressing need for high-quality data in model training. Their vision is creating flexible AI solutions that truly adapt to enterprise workflows.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
ANECDOTE

Mark's Career Transition

  • Mark Huang's career path transitioned from quantitative finance to Silicon Valley.
  • He worked at Box and Splunk before co-founding Gradient.
INSIGHT

Gradient AI's Vision

  • Gradient AI aims to enable businesses to transition from rule-based automation to autonomous agent workflows.
  • They built a horizontal platform to handle out-of-domain problems and improve alongside users.
INSIGHT

Long Context and Llama 3

  • Long context allows for better out-of-distribution generalization in AI.
  • Gradient chose Llama 3 because of its short context length and criticisms surrounding it.
Get the Snipd Podcast app to discover more snips from this episode
Get the app