The Infra Pod

Building the Future of AI with Long-term Memory (Chat with Charles from Letta)

Aug 11, 2025
Dive into the world of AI memory and discover its revolutionary potential. Charles from Letta discusses how long-term memory can enhance productivity tools and coding assistants. The concept of 'sleep time compute' is explored, highlighting continuous AI efficiency. Future trends suggest shared memory could outperform models in importance, shaping the way software interacts with users and improves collaboration. Tune in to learn about the early stages of AI memory technology and the challenges that lie ahead in creating more effective systems.
Ask episode
AI Snips
Chapters
Transcript
Episode notes
INSIGHT

Sleep-Time Compute Lets Agents Think Offline

  • Sleep-time compute runs models asynchronously while users are away to precompute intelligence.
  • Charles stresses this only pays off if the agent has durable memory to store and reuse those offline insights.
ADVICE

Precompute Code Insights During Idle Time

  • Let agents inspect repos and logs during idle time to build mental maps and deep wikis.
  • Precompute reflections and write patterns to memory so future coding sessions produce much better contextual answers.
INSIGHT

Shared Memory Should Be Human-Readable Text

  • Charles promotes shared memory stored as plain text (e.g., Postgres rows) so agents and humans can read and edit it.
  • He predicts many productivity agents will read from common memory segments rather than isolated context windows.
Get the Snipd Podcast app to discover more snips from this episode
Get the app