Deep Papers

Hungry Hungry Hippos - H3

Feb 13, 2023
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

FFT Enables Efficient Long Convolutions

  • FFT lets SSMs compute very long convolutions in near-linear time. That algorithmic property explains why SSMs scale much better than attention for long inputs.
ANECDOTE

Audio Shows SSM Strength

  • Audio waveforms are a clear example where transformers choke due to sequence length (e.g., 64k samples/sec). SSMs can process such sequences and model waveform structure effectively.
ANECDOTE

Why 'Hungry Hungry Hippos'?

  • The 'Hippo' name came from 'hippocampus' and earlier lab work on memory mechanisms. 'Hungry Hungry Hippos' playfully signals the dual-SSM design.
Get the Snipd Podcast app to discover more snips from this episode
Get the app