

Long Context Language Models and their Biological Applications with Eric Nguyen - #690
9 snips Jun 25, 2024
Eric Nguyen, a PhD student at Stanford, dives deep into his research on long context foundation models, specifically Hyena and its applications in biology. He explains the limitations of traditional transformers in processing lengthy sequences and how convolutional models provide innovative solutions. Nguyen introduces Hyena DNA, designed for long-range DNA dependencies, and discusses Evo, a hybrid model with massive parameters for DNA generation. The podcast touches on exciting applications in CRISPR gene editing and the implications of using AI in biological research.
AI Snips
Chapters
Transcript
Episode notes
Transformer Limitations
- Transformers struggle with long sequences due to quadratic time complexity.
- Doubling sequence length quadruples computation, making them inefficient for very long sequences like DNA.
Hyena Architecture
- Hyena uses long convolutions with implicit kernels for parameter efficiency.
- It leverages a fast Fourier transform (FFT) for near-linear time complexity with long sequences.
Explainability in Hyena
- While state-space models offer explainability through explicit state evolution, Hyena prioritizes scalability.
- Hyena uses "attention map-like" visualizations of convolutions for interpretability.