The Missing Architecture Behind Autonomous AI with Jacob Buckman

Embed Block
Add an embed URL or code. Learn more

In this episode of Machine Minds, we step beyond today’s transformer-dominated AI landscape and into a deeper conversation about what’s missing on the path to truly autonomous, long-horizon intelligence. Jacob Buckman, co-founder and CEO of Manifest AI, joins Greg to explore why current AI systems struggle with long-term reasoning, persistent memory, and extended task execution—and what it will take to unlock the next paradigm.

Jacob’s journey into AI began early, fueled by science fiction, programming, and a fascination with building systems that could do meaningful work autonomously. From studying and conducting research at Carnegie Mellon to working at Google Brain, he watched deep learning unify once-fragmented AI subfields—vision, language, speech—under a single scalable framework. That unification shaped his conviction that the next breakthrough wouldn’t come from incremental tuning, but from rethinking a fundamental architectural bottleneck.

At Manifest AI, Jacob and his team are tackling what they believe is the missing piece: scalable long-context intelligence. Their work centers on replacing transformer attention with a new family of architectures called retention models, designed to compress and retain relevant information over time—rather than repeatedly replaying massive histories. The goal: AI systems that can reason, learn, and work continuously over hours, days, or longer.

In this conversation, Greg and Jacob explore:

  • Jacob’s path from aspiring scientist to AI researcher and founder—and why curiosity plus first principles thinking matter more than trends

  • Why today’s large language models excel at short tasks but break down over long horizons

  • The core limitation of transformer attention—and why “attention is all you need” may no longer hold

  • How retention architectures unify the strengths of transformers and recurrent neural networks

  • What it means for an AI system to compress knowledge instead of endlessly appending memory

  • Why long-term reasoning, iterative problem solving, and true autonomy require architectural change—not orchestration hacks

  • The misconception that agent orchestration can substitute for unified, persistent intelligence

  • How long-context models could reshape agents from short-lived “consultants” into persistent, personalized collaborators

  • The technical challenge of translating theoretical breakthroughs into high-performance GPU kernels

  • Why Manifest AI is open source—and how their work aims to move the entire field forward

  • Lessons from unifying AI subfields, the “bitter lesson” of scale, and avoiding ad-hoc solutions that won’t last

  • Jacob’s view on cost, intelligence density, and why better architectures will increase—not reduce—investment in AI

  • Advice for founders and researchers: focus relentlessly on the single bottleneck that matters most

If you’re building AI systems, researching foundations of intelligence, or trying to understand what comes after today’s models, this episode offers a rare, deeply reasoned look at where the field may be heading—and why architectural simplicity could unlock far more than brute force scale.

Learn more about Manifest AI: https://manifestai.com

Explore the open-source retention models: pip install retention

Connect with Jacob Buckman on LinkedIn: https://www.linkedin.com/in/jacobbuckman

Connect with Greg Toroosian on Linkedin: https://www.linkedin.com/in/gregtoroosian

Next
Next

Fixing the Last Manual Step in Modern Logistics with Chris Smith