The TWIML AI Podcast (formerly This Week in Machine Learning & Artificial Intelligence)

Long Context Language Models and their Biological Applications with Eric Nguyen - #690

9 snips
Jun 25, 2024
Eric Nguyen, a PhD student at Stanford, dives deep into his research on long context foundation models, specifically Hyena and its applications in biology. He explains the limitations of traditional transformers in processing lengthy sequences and how convolutional models provide innovative solutions. Nguyen introduces Hyena DNA, designed for long-range DNA dependencies, and discusses Evo, a hybrid model with massive parameters for DNA generation. The podcast touches on exciting applications in CRISPR gene editing and the implications of using AI in biological research.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Transformer Limitations

  • Transformers struggle with long sequences due to quadratic time complexity.
  • Doubling sequence length quadruples computation, making them inefficient for very long sequences like DNA.
INSIGHT

Hyena Architecture

  • Hyena uses long convolutions with implicit kernels for parameter efficiency.
  • It leverages a fast Fourier transform (FFT) for near-linear time complexity with long sequences.
INSIGHT

Explainability in Hyena

  • While state-space models offer explainability through explicit state evolution, Hyena prioritizes scalability.
  • Hyena uses "attention map-like" visualizations of convolutions for interpretability.
Get the Snipd Podcast app to discover more snips from this episode
Get the app