Machine Minds
Machine Minds - the minds behind the machines! This is the show where we dive deep into the intricate worlds of robotics, AI, and Hard Tech. In each episode, we bring you intimate conversations with the founders, investors, and trailblazers who are at the heart of these tech revolutions. We dig into their journeys, the challenges they've overcome, and the breakthroughs that are shaping our future. Join us as we explore how these machine minds are transforming the way we live, work, and understand our world.
Machine Minds
The Missing Architecture Behind Autonomous AI with Jacob Buckman
In this episode of Machine Minds, we step beyond today’s transformer-dominated AI landscape and into a deeper conversation about what’s missing on the path to truly autonomous, long-horizon intelligence. Jacob Buckman, co-founder and CEO of Manifest AI, joins Greg to explore why current AI systems struggle with long-term reasoning, persistent memory, and extended task execution—and what it will take to unlock the next paradigm.
Jacob’s journey into AI began early, fueled by science fiction, programming, and a fascination with building systems that could do meaningful work autonomously. From studying and conducting research at Carnegie Mellon to working at Google Brain, he watched deep learning unify once-fragmented AI subfields—vision, language, speech—under a single scalable framework. That unification shaped his conviction that the next breakthrough wouldn’t come from incremental tuning, but from rethinking a fundamental architectural bottleneck.
At Manifest AI, Jacob and his team are tackling what they believe is the missing piece: scalable long-context intelligence. Their work centers on replacing transformer attention with a new family of architectures called retention models, designed to compress and retain relevant information over time—rather than repeatedly replaying massive histories. The goal: AI systems that can reason, learn, and work continuously over hours, days, or longer.
In this conversation, Greg and Jacob explore:
- Jacob’s path from aspiring scientist to AI researcher and founder—and why curiosity plus first principles thinking matter more than trends
- Why today’s large language models excel at short tasks but break down over long horizons
- The core limitation of transformer attention—and why “attention is all you need” may no longer hold
- How retention architectures unify the strengths of transformers and recurrent neural networks
- What it means for an AI system to compress knowledge instead of endlessly appending memory
- Why long-term reasoning, iterative problem solving, and true autonomy require architectural change—not orchestration hacks
- The misconception that agent orchestration can substitute for unified, persistent intelligence
- How long-context models could reshape agents from short-lived “consultants” into persistent, personalized collaborators
- The technical challenge of translating theoretical breakthroughs into high-performance GPU kernels
- Why Manifest AI is open source—and how their work aims to move the entire field forward
- Lessons from unifying AI subfields, the “bitter lesson” of scale, and avoiding ad-hoc solutions that won’t last
- Jacob’s view on cost, intelligence density, and why better architectures will increase—not reduce—investment in AI
- Advice for founders and researchers: focus relentlessly on the single bottleneck that matters most
If you’re building AI systems, researching foundations of intelligence, or trying to understand what comes after today’s models, this episode offers a rare, deeply reasoned look at where the field may be heading—and why architectural simplicity could unlock far more than brute force scale.
Learn more about Manifest AI: https://manifestai.com
Explore the open-source retention models: pip install retention
Connect with Jacob Buckman on LinkedIn: https://www.linkedin.com/in/jacobbuckman
Connect with Greg Toroosian on Linkedin: https://www.linkedin.com/in/gregtoroosian