#304 Matt Zeiler: Why Government And Enterprises Choose Clarifai For AI Ops
Release Date: 11/28/2025
Eye On A.I.
Inference is now the biggest challenge in enterprise AI. In this episode of Eye on AI, Craig Smith speaks with Nick Pandher, VP of Product at Cirrascale, about why AI is shifting from model training to inference at scale. As AI moves into production, enterprises are prioritizing performance, latency, reliability, and cost efficiency over raw compute. The conversation covers the rise of inference-first infrastructure, the limits of hyperscalers, the emergence of neoclouds, and how agentic AI is driving always-on inference workloads. Nick also explains how inference-optimized hardware and...
info_outlineEye On A.I.
In this episode of Eye on AI, we sit down with Evan Reiser, co-founder and CEO of Abnormal AI, to unpack how AI has fundamentally changed the cybersecurity landscape. We explore why social engineering remains the most costly form of cybercrime, how generative AI has lowered the barrier for sophisticated attacks, and why humans have become the primary attack surface in modern security. Evan explains why traditional, signature-based defenses fall short, how behavioral AI detects threats that have never existed before, and what it means to build security systems that understand how people...
info_outlineEye On A.I.
In this episode of Eye on AI, Craig Smith speaks with Jonathan Wall, founder and CEO of Runloop AI, about why AI agents require an entirely new approach to compute infrastructure. Jonathan explains why agents behave very differently from traditional servers, why giving agents their own isolated computers unlocks new capabilities, and how agent-native infrastructure is emerging as a critical layer of the AI stack. The conversation also covers scaling agents in production, building trust through benchmarking and human-in-the-loop workflows, and what agent-driven systems mean for the...
info_outlineEye On A.I.
In this episode of Eye on AI, Craig Smith sits down with Anurag Dhingra, Senior Vice President and General Manager at Cisco, to explore where AI is actually creating value inside the enterprise. Rather than focusing on flashy demos or speculative futures, this conversation goes deep into the invisible layer powering modern AI: infrastructure. Anurag breaks down how AI is being embedded into enterprise networking, security, observability, and collaboration systems to solve real operational problems at scale. From self-healing networks and agentic AI to edge computing, robotics, and...
info_outlineEye On A.I.
This episode is sponsored by AGNTCY. Unlock agents at scale with an open Internet of Agents. Visit and add your support. Most large language models today generate text one token at a time. That design choice creates a hard limit on speed, cost, and scalability. In this episode of Eye on AI, Stefano Ermon breaks down diffusion language models and why a parallel, inference-first approach could define the next generation of LLMs. We explore how diffusion models differ from autoregressive systems, why inference efficiency matters more than training scale, and what this shift means for...
info_outlineEye On A.I.
This episode is sponsored by AGNTCY. Unlock agents at scale with an open Internet of Agents. Visit and add your support. Why are AI, biotechnology, and gene editing converging right now, and what does that mean for the future of humanity? In this episode of Eye on AI, host Craig Smith sits down with futurist and author Jamie Metzl to explore the superconvergence of artificial intelligence, genomics, and exponential technologies that are reshaping life on Earth. We examine the ethical and scientific realities behind human genome editing, the controversy around CRISPR babies, and why...
info_outlineEye On A.I.
Try OCI for free at This episode is sponsored by Oracle. OCI is the next-generation cloud designed for every workload – where you can run any application, including any AI projects, faster and more securely for less. On average, OCI costs 50% less for compute, 70% less for storage, and 80% less for networking. Join Modal, Skydance Animation, and today’s innovative AI tech companies who upgraded to OCI…and saved. Why is AI moving from the cloud to our devices, and what makes on device intelligence finally practical at scale? In this episode of Eye on AI, host Craig Smith...
info_outlineEye On A.I.
This episode is sponsored by AGNTCY. Unlock agents at scale with an open Internet of Agents. Visit and add your support. Why is AI so powerful in the cloud but still so limited inside everyday devices, and what would it take to run intelligent systems locally without draining battery or sacrificing privacy? In this episode of Eye on AI, host Craig Smith speaks with Steve Brightfield, Chief Marketing Officer at BrainChip, about neuromorphic computing and why brain inspired architectures may be the key to the future of edge AI. We explore how neuromorphic systems differ from traditional...
info_outlineEye On A.I.
This episode is sponsored by AGNTCY. Unlock agents at scale with an open Internet of Agents. Visit and add your support. Why do some AI agents attempt to bypass shutdown, and what does this behavior reveal about the future of AI safety? In this episode of Eye on AI, host Craig Smith speaks with Jeffrey Ladish of Palisade Research to examine what recent shutdown experiments with agentic LLMs tell us about control, alignment, and the real world limits of current guardrails. We explore how models behave when placed in virtual machine environments, why some agents edit or disable...
info_outlineEye On A.I.
Why are enterprises struggling to turn AI hype into real workplace transformation, and how is Lenovo using agentic AI to actually close that gap? In this episode of Eye on AI, host Craig Smith talks with Rakshit Ghura about how his team is reinventing the modern workplace with an omnichannel AI architecture powered by a fleet of specialized agents. We explore how Lenovo has evolved from a hardware company into a global solutions provider, and how its Care of One platform uses persona based design to improve employee experience, reduce downtime, and personalize support across IT, HR, and...
info_outlineTry OCI for free at http://oracle.com/eyeonai
This episode is sponsored by Oracle. OCI is the next-generation cloud designed for every workload – where you can run any application, including any AI projects, faster and more securely for less. On average, OCI costs 50% less for compute, 70% less for storage, and 80% less for networking.
Join Modal, Skydance Animation, and today’s innovative AI tech companies who upgraded to OCI…and saved.
Why is AI inference becoming the new battleground for speed, cost, and real world scalability, and how are companies like Clarifai reshaping the AI stack by optimizing every token and every deployment?
In this episode of Eye on AI, host Craig Smith sits down with Clarifai founder and CEO Matt Zeiler to explore why inference is now more important than training and how a unified compute orchestration layer is changing the way teams run LLMs and agentic systems.
We look at what makes high performance inference possible across cloud, on prem, and edge environments, how to get faster responses from large language models, and how to cut GPU spend without sacrificing intelligence or accuracy. Learn how organizations operate AI systems in regulated industries, how government teams and enterprises use Clarifai to deploy models securely, and which bottlenecks matter most when running long context, multimodal, or high throughput applications.
You will also hear how to optimize your own AI workloads with better token throughput, how to choose the right hardware strategy for scale, and how inference first architecture can turn models into real products. This conversation breaks down the tools, techniques, and design patterns that can help your AI agents run faster, cheaper, and more reliably in production.
Stay Updated:
Craig Smith on X: https://x.com/craigss
Eye on A.I. on X: https://x.com/EyeOn_AI