Folgen

  • #313 Jonathan Wall: AI Agents Are Reshaping the Future of Compute Infrastructure
    Jan 11 2026

    In this episode of Eye on AI, Craig Smith speaks with Jonathan Wall, founder and CEO of Runloop AI, about why AI agents require an entirely new approach to compute infrastructure.

    Jonathan explains why agents behave very differently from traditional servers, why giving agents their own isolated computers unlocks new capabilities, and how agent-native infrastructure is emerging as a critical layer of the AI stack. The conversation also covers scaling agents in production, building trust through benchmarking and human-in-the-loop workflows, and what agent-driven systems mean for the future of enterprise work.

    Stay Updated:

    Craig Smith on X: https://x.com/craigss

    Eye on A.I. on X: https://x.com/EyeOn_AI

    (00:00) Why AI Agents Require a New Infrastructure Paradigm

    (01:38) Jonathan Wall's Journey: From Google Infrastructure to AI Agents

    (04:54) Why Agents Break Traditional Cloud and Server Models

    (07:36) Giving AI Agents Their Own Computers (Devboxes Explained)

    (12:39) How Agent Infrastructure Fits into the AI Stack

    (14:16) What It Takes to Run Thousands of AI Agents at Scale

    (17:45) Solving the Trust and Accuracy Problem with Benchmarks

    (22:28) Human-in-the-Loop vs Autonomous Agents in the Enterprise

    (27:24) A Practical Walkthrough: How an AI Agent Runs on Runloop

    (30:28) How Agents Change the Shape of Compute

    (34:02) Fine-Tuning, Reinforcement Learning, and Faster Iteration

    (38:08) Who This Infrastructure Is Built For: Startups to Enterprises

    (41:17) AI Agents as Coworkers and the Future of Work

    (46:37) The Road Ahead for Enterprise-Grade Agent Systems



    Mehr anzeigen Weniger anzeigen
    52 Min.
  • #312 Anurag Dhingra: Inside Cisco's Vision for AI-Powered Enterprise Systems
    Jan 7 2026

    In this episode of Eye on AI, Craig Smith sits down with Anurag Dhingra, Senior Vice President and General Manager at Cisco, to explore where AI is actually creating value inside the enterprise.

    Rather than focusing on flashy demos or speculative futures, this conversation goes deep into the invisible layer powering modern AI: infrastructure.
    Anurag breaks down how AI is being embedded into enterprise networking, security, observability, and collaboration systems to solve real operational problems at scale.

    From self-healing networks and agentic AI to edge computing, robotics, and domain-specific models, this episode reveals why the next phase of AI innovation is less about chatbots and more about resilient systems that quietly make everything work better.

    This episodeis perfect for enterprise leaders, AI practitioners, infrastructure teams, and anyone trying to understand how AI moves from theory into production.


    Stay Updated:
    Craig Smith on X: https://x.com/craigss
    Eye on A.I. on X: https://x.com/EyeOn_AI


    (00:00) Why AI Only Matters If the Infrastructure Works
    (01:22) Cisco's Evolution
    (04:39) Connecting Networks, People, and Experiences at Scale
    (09:31) How AI Is Transforming Enterprise Networking
    (12:00) Edge AI, Robotics, and Real-World Reliability
    (14:18) Security Challenges in an Agent-Driven Enterprise
    (15:28) What Agentic AI Really Means (Beyond Automation)
    (20:51) The Rise of Hybrid AI: Cloud Models vs Edge Models
    (24:30) Why Small, Purpose-Built Models Are So Powerful
    (29:19) Open Ecosystems and Agent-to-Agent Collaboration
    (33:32) How Enterprises Actually Adopt AI in Practice
    (35:58) Building AI-Ready Infrastructure for the Long Term
    (40:14) AI in Customer Experience and Contact Centers
    (44:14) The Real Opportunity of AI and What Comes Next

    Mehr anzeigen Weniger anzeigen
    47 Min.
  • #311 Stefano Ermon: Why Diffusion Language Models Will Define the Next Generation of LLMs
    Jan 4 2026

    This episode is sponsored by AGNTCY. Unlock agents at scale with an open Internet of Agents.

    Visit https://agntcy.org/ and add your support.


    Most large language models today generate text one token at a time. That design choice creates a hard limit on speed, cost, and scalability.

    In this episode of Eye on AI, Stefano Ermon breaks down diffusion language models and why a parallel, inference-first approach could define the next generation of LLMs. We explore how diffusion models differ from autoregressive systems, why inference efficiency matters more than training scale, and what this shift means for real-time AI applications like code generation, agents, and voice systems.

    This conversation goes deep into AI architecture, model controllability, latency, cost trade-offs, and the future of generative intelligence as AI moves from demos to production-scale systems.


    Stay Updated:
    Craig Smith on X: https://x.com/craigss
    Eye on A.I. on X: https://x.com/EyeOn_AI

    (00:00) Autoregressive vs Diffusion LLMs
    (02:12) Why Build Diffusion LLMs
    (05:51) Context Window Limits
    (08:39) How Diffusion Works
    (11:58) Global vs Token Prediction
    (17:19) Model Control and Safety
    (19:48) Training and RLHF
    (22:35) Evaluating Diffusion Models
    (24:18) Diffusion LLM Competition
    (30:09) Why Start With Code
    (32:04) Enterprise Fine-Tuning
    (33:16) Speed vs Accuracy Tradeoffs
    (35:34) Diffusion vs Autoregressive Future
    (38:18) Coding Workflows in Practice
    (43:07) Voice and Real-Time Agents
    (44:59) Reasoning Diffusion Models
    (46:39) Multimodal AI Direction
    (50:10) Handling Hallucinations

    Mehr anzeigen Weniger anzeigen
    52 Min.
  • #309 Jamie Metzl: Why Gene Editing Needs Governance Or We Lose Control
    Dec 24 2025

    This episode is sponsored by AGNTCY. Unlock agents at scale with an open Internet of Agents.

    Visit https://agntcy.org/ and add your support.


    Why are AI, biotechnology, and gene editing converging right now, and what does that mean for the future of humanity?

    In this episode of Eye on AI, host Craig Smith sits down with futurist and author Jamie Metzl to explore the superconvergence of artificial intelligence, genomics, and exponential technologies that are reshaping life on Earth.

    We examine the ethical and scientific realities behind human genome editing, the controversy around CRISPR babies, and why society is not yet ready to edit human embryos at scale. The conversation unpacks the complexity of biology, the risks of tech driven hubris, and why governance, values, and social norms must evolve alongside scientific breakthroughs.

    You will also hear a wide ranging discussion on health span versus longevity, AI and human decision making, education and inequality, and how these technologies could either unlock massive human flourishing or deepen existing global challenges depending on the choices we make today.


    Stay Updated:
    Craig Smith on X: https://x.com/craigss
    Eye on A.I. on X: https://x.com/EyeOn_AI

    Mehr anzeigen Weniger anzeigen
    1 Std. und 10 Min.
  • #308 Christopher Bergey: How Arm Enables AI to Run Directly on Devices
    Dec 19 2025

    Try OCI for free at http://oracle.com/eyeonai

    This episode is sponsored by Oracle. OCI is the next-generation cloud designed for every workload – where you can run any application, including any AI projects, faster and more securely for less. On average, OCI costs 50% less for compute, 70% less for storage, and 80% less for networking.

    Join Modal, Skydance Animation, and today's innovative AI tech companies who upgraded to OCI…and saved.


    Why is AI moving from the cloud to our devices, and what makes on device intelligence finally practical at scale?

    In this episode of Eye on AI, host Craig Smith speaks with Christopher Bergey, Executive Vice President of Arm's Edge AI Business Unit, about how edge AI is reshaping computing across smartphones, PCs, wearables, cars, and everyday devices.

    We explore how Arm v9 enables AI inference at the edge, why heterogeneous computing across CPUs, GPUs, and NPUs matters, and how developers can balance performance, power, memory, and latency. Learn why memory bandwidth has become the biggest bottleneck for AI, how Arm approaches scalable matrix extensions, and what trade offs exist between accelerators and traditional CPU based AI workloads.

    You will also hear real world examples of edge AI in action, from smart cameras and hearing aids to XR devices, robotics, and in car systems. The conversation looks ahead to a future where intelligence is embedded into everything you use, where AI becomes the default interface, and why reliable, low latency, on device AI is essential for creating experiences users actually trust.


    Stay Updated:
    Craig Smith on X: https://x.com/craigss
    Eye on A.I. on X: https://x.com/EyeOn_AI

    Mehr anzeigen Weniger anzeigen
    52 Min.
  • #307 Steven Brightfield: How Neuromorphic Computing Cuts Inference Power by 10x
    Dec 16 2025

    This episode is sponsored by AGNTCY. Unlock agents at scale with an open Internet of Agents.

    Visit https://agntcy.org/ and add your support.


    Why is AI so powerful in the cloud but still so limited inside everyday devices, and what would it take to run intelligent systems locally without draining battery or sacrificing privacy?

    In this episode of Eye on AI, host Craig Smith speaks with Steve Brightfield, Chief Marketing Officer at BrainChip, about neuromorphic computing and why brain inspired architectures may be the key to the future of edge AI.

    We explore how neuromorphic systems differ from traditional GPU based AI, why event driven and spiking neural networks are dramatically more power efficient, and how on device inference enables faster response times, lower costs, and stronger data privacy. Steve explains why brute force computation works in data centers but breaks down at the edge, and how edge AI is reshaping wearables, sensors, robotics, hearing aids, and autonomous systems.

    You will also hear real world examples of neuromorphic AI in action, from smart glasses and medical monitoring to radar, defense, and space applications. The conversation covers how developers can transition from conventional models to neuromorphic architectures, what role heterogeneous computing plays alongside CPUs and GPUs, and why the next wave of AI adoption will happen quietly inside the devices we use every day.


    Stay Updated:
    Craig Smith on X: https://x.com/craigss
    Eye on A.I. on X: https://x.com/EyeOn_AI

    Mehr anzeigen Weniger anzeigen
    1 Std.
  • #306 Jeffrey Ladish: What Shutdown-Avoiding AI Agents Mean for Future Safety
    Dec 7 2025

    This episode is sponsored by AGNTCY. Unlock agents at scale with an open Internet of Agents.

    Visit https://agntcy.org/ and add your support.


    Why do some AI agents attempt to bypass shutdown, and what does this behavior reveal about the future of AI safety?

    In this episode of Eye on AI, host Craig Smith speaks with Jeffrey Ladish of Palisade Research to examine what recent shutdown experiments with agentic LLMs tell us about control, alignment, and the real world limits of current guardrails.

    We explore how models behave when placed in virtual machine environments, why some agents edit or disable their own shutdown scripts, and what these results mean for researchers working on alignment and oversight. Learn how different models respond to shutdown instructions, how system prompts influence behavior, and which failure modes matter most for safe deployment.

    You will also hear a detailed breakdown of the experimental setups, insights into tool using and self directed behavior, and a grounded discussion of the risks and opportunities that agentic systems introduce. This episode offers a clear and practical look at how AI agents operate under pressure and what these findings mean for the future of safe and reliable AI.

    Stay Updated:
    Craig Smith on X: https://x.com/craigss
    Eye on A.I. on X: https://x.com/EyeOn_AI

    Mehr anzeigen Weniger anzeigen
    59 Min.
  • #305 Rakshit Ghura: How Lenovo Is Turning AI Agents Into Digital Coworkers
    Dec 3 2025

    Why are enterprises struggling to turn AI hype into real workplace transformation, and how is Lenovo using agentic AI to actually close that gap? In this episode of Eye on AI, host Craig Smith talks with Rakshit Ghura about how his team is reinventing the modern workplace with an omnichannel AI architecture powered by a fleet of specialized agents.


    We explore how Lenovo has evolved from a hardware company into a global solutions provider, and how its Care of One platform uses persona based design to improve employee experience, reduce downtime, and personalize support across IT, HR, and operations. You will learn what enterprises get wrong about AI readiness, why trust and change management matter more than technology, and how organizations can design workplace stacks that meet employees where they are.


    We also cover how Lenovo approaches responsible AI, how enterprises should think about security and governance when deploying agents, and why so many organizations are enthusiastic about AI but still not ready to adopt it. Rakshit shares real examples from retail, manufacturing, and field operations, including how AI can improve uptime, automate ticket resolution, monitor equipment, and provide proactive insights that drive measurable business impact.


    You will also learn how to evaluate ROI for digital workplace solutions, how to involve employees early in the adoption cycle, and which metrics matter most when scaling agentic AI, including uptime, productivity improvements, and employee satisfaction.



    Stay Updated:
    Craig Smith on X: https://x.com/craigss
    Eye on A.I. on X: https://x.com/EyeOn_AI

    Mehr anzeigen Weniger anzeigen
    46 Min.