loader from loading.io

d-Matrix - Ultra-low Latency Batched Inference for Gen AI

Tech Talks Daily

Release Date: 03/07/2026

How Phenom Is Using AI To Transform Hiring And Talent Intelligence show art How Phenom Is Using AI To Transform Hiring And Talent Intelligence

Tech Talks Daily

How can organizations use AI to transform hiring while still protecting the human element at the heart of work? In this episode of Tech Talks Daily, I sit down with Mahe Bayireddi, co-founder and CEO of Phenom, to explore how artificial intelligence is reshaping the way companies attract, hire, and develop talent.  Our conversation comes at an interesting moment for the company, following the announcement that Phenom has acquired Be Applied, an AI-driven cognitive assessment platform designed to validate candidate and employee capabilities at scale. The move follows an earlier acquisition...

info_outline
How CISOs Can Earn Real Influence In The Boardroom With Rapid7 show art How CISOs Can Earn Real Influence In The Boardroom With Rapid7

Tech Talks Daily

How does a CISO turn cybersecurity from a technical conversation into a business conversation that boards actually care about? In this episode of Tech Talks Daily, I sit down with Thom Langford, EMEA CTO at Rapid7 and a former CISO, to explore what he calls the second phase of cybersecurity leadership. For years, the industry worked hard to secure a seat at the boardroom table. In many organizations, that mission has largely succeeded. But as Thom explains, gaining access was only the first step. The real challenge now is communicating security in a way that drives meaningful business...

info_outline
How Shokz Is Leading The Rise Of Open-Ear Headphones show art How Shokz Is Leading The Rise Of Open-Ear Headphones

Tech Talks Daily

What if the next big shift in personal audio is not about blocking the world out, but staying connected to it? In this episode of Tech Talks Daily, I sit down with Nicole from Shokz to talk about why open-ear headphones are suddenly everywhere, and why this category is moving from niche curiosity to everyday essential. For years, the audio market was obsessed with sealing users off from the outside world. Now the conversation is changing. More people want to hear their music, podcasts, and calls without losing awareness of traffic, fellow commuters, colleagues, or the world happening...

info_outline
d-Matrix - Ultra-low Latency Batched Inference for Gen AI  show art d-Matrix - Ultra-low Latency Batched Inference for Gen AI

Tech Talks Daily

What happens when the real bottleneck in artificial intelligence is no longer training models, but actually running them at scale? In this episode of Tech Talks Daily, I sit down with Satyam Srivastava from d-Matrix to explore a shift that is quietly reshaping the entire AI infrastructure landscape. While much of the early AI race focused on training ever larger models, the next phase of AI adoption is increasingly defined by inference. That is the moment when trained models are deployed and used to generate real-world results millions of times a day. Satyam brings a unique perspective shaped...

info_outline
How Scale Computing Is Powering The Next Wave Of Edge Infrastructure show art How Scale Computing Is Powering The Next Wave Of Edge Infrastructure

Tech Talks Daily

How should businesses rethink infrastructure when applications, data, and users are increasingly spread across thousands of locations? In this episode of Tech Talks Daily, I sit down with Mark Cree, President and Chief Operating Officer at Scale Computing, to talk about why the future of enterprise infrastructure is moving closer to where data is actually created. This conversation was recorded following the 66th edition of The IT Press Tour, where some of the most interesting conversations in enterprise infrastructure centered on what happens when businesses move away from oversized,...

info_outline
How InfoScale Is Redefining Enterprise Resilience In A Multi-Cloud World show art How InfoScale Is Redefining Enterprise Resilience In A Multi-Cloud World

Tech Talks Daily

How confident are you that your business could recover from a cyberattack, cloud outage, or infrastructure failure in minutes rather than hours or even days? In this episode of Tech Talks Daily, I explore the changing nature of enterprise resilience with Joseph D'Angelo and Cassie Stanek from InfoScale, now part of Cloud Software Group. Our conversation looks at why many organizations still rely on backup and replication strategies that were designed for a very different era of IT. In a world of hybrid infrastructure, multi-cloud deployments, and increasingly complex application...

info_outline
How Ticket Fairy Is Rebuilding The Technology Behind Live Events show art How Ticket Fairy Is Rebuilding The Technology Behind Live Events

Tech Talks Daily

Have you ever bought a ticket to a show and wondered why the experience still feels strangely disconnected, with one app for ticketing, another for marketing, another for refunds, and a dozen spreadsheets held together by late nights and good intentions? In this episode of Tech Talks Daily, I’m joined by Ritesh Patel, co-founder of Ticket Fairy, to talk about the technology behind live events and why it has lagged behind other industries in some surprisingly familiar ways. Ritesh makes the case that most organizers are operating more like creative founders than corporate operators, building...

info_outline
Hiring AI Talent Across Borders With Alcor  show art Hiring AI Talent Across Borders With Alcor

Tech Talks Daily

Have you ever looked at a global hiring plan and wondered whether you are building a team, or accidentally buying a bundle of hidden fees, legal risk, and avoidable stress? In this episode, I’m joined by Oksana Petrus from Alcor, where she leads customer success and operations, helping tech companies build and scale engineering teams across Eastern Europe and Latin America. If you have ever tried to expand beyond your home market, you know the promise is real, access to great talent, broader coverage across time zones, and the chance to build faster. But the reality can get messy quickly...

info_outline
How Flashfood Uses Data And AI To Solve The Grocery Food Waste Crisis show art How Flashfood Uses Data And AI To Solve The Grocery Food Waste Crisis

Tech Talks Daily

How can a world that produces more than enough food still leave millions of people struggling to put a healthy meal on the table? In this episode of Tech Talks Daily, I speak with Jordan Schenck, CEO of Flashfood, about the growing paradox at the heart of our global food system. Grocery prices are climbing, families everywhere are making harder choices at the checkout, and food banks are seeing rising demand. Yet at the same time, vast quantities of perfectly edible food never make it onto a plate. Jordan shares the startling scale of the problem. In North America alone, billions of pounds of...

info_outline
SmartRecruiters On Turning AI Experiments Into Business Outcomes show art SmartRecruiters On Turning AI Experiments Into Business Outcomes

Tech Talks Daily

Is 2026 the year AI finally has to prove it is worth the investment? In this episode, I’m joined by Chris Riche-Webber, VP of Business Intelligence and Analytics at SmartRecruiters, to explore why so many AI and agentic AI initiatives stall after the pilot phase and what separates the projects that scale from the ones that quietly disappear. With Gartner predicting that more than 40 percent of agentic AI programs could be cancelled by 2027, Chris brings a pragmatic, data-led perspective on what is really happening inside organizations as the hype meets operational reality. We talk about the...

info_outline
 
More Episodes

What happens when the real bottleneck in artificial intelligence is no longer training models, but actually running them at scale?

In this episode of Tech Talks Daily, I sit down with Satyam Srivastava from d-Matrix to explore a shift that is quietly reshaping the entire AI infrastructure landscape. While much of the early AI race focused on training ever larger models, the next phase of AI adoption is increasingly defined by inference. That is the moment when trained models are deployed and used to generate real-world results millions of times a day.

Satyam brings a unique perspective shaped by years of experience in signal processing, machine learning, and hardware architecture, including time spent at NVIDIA and Intel working on graphics, media technologies, and AI systems. Now at d-Matrix, he is helping design next-generation computing architectures focused on one of the biggest challenges facing the AI industry today: efficiently running large language models without overwhelming data centers with unsustainable power and infrastructure demands.

During our conversation, we explored why the industry underestimated the infrastructure implications of inference at scale. While training large models grabs headlines, the real operational pressure often comes later when those models must serve millions of queries in real time. That shift places enormous strain on memory bandwidth, energy consumption, and data movement inside modern data centers.

Satyam explains how d-Matrix identified this challenge years before generative AI exploded into the mainstream. Instead of focusing on training hardware like many AI startups at the time, the company concentrated on inference efficiency. That decision is becoming increasingly relevant as organizations begin to realize that simply adding more GPUs to data centers is not a sustainable long-term strategy.

We also discuss the growing power constraints surrounding AI infrastructure, and why efficiency-driven design may be the only realistic path forward. With electricity supply, cooling capacity, and semiconductor availability all becoming limiting factors, the industry is being forced to rethink how AI systems are architected. Custom silicon, purpose-built accelerators, and heterogeneous computing environments are now emerging as key pieces of the puzzle.

The conversation also touches on the geopolitical and economic importance of AI semiconductor leadership, and why the relationship between frontier AI labs, infrastructure providers, and chip designers is becoming increasingly strategic. As governments and companies compete to maintain technological leadership, the question of who controls the hardware powering AI may prove just as important as the models themselves.

Looking ahead, Satyam shares his perspective on how the role of engineers will evolve as AI infrastructure becomes more specialized and energy-aware. Foundational engineering skills remain essential, but the next generation of engineers will also need to think in terms of entire systems, combining software, hardware, and AI tools to build more efficient computing environments.

As AI continues to move from research labs into everyday products and services, are organizations prepared for the infrastructure shift that comes with an inference-driven future? And could efficiency, rather than raw computing power, become the defining metric of the next phase of the AI race?