loader from loading.io

Heather Domin: From Principles to Practice

The Road to Accountable AI

Release Date: 10/02/2025

Katie Fowler (Thompson Reuters Foundation): How 3,000 Companies Approach AI Governance show art Katie Fowler (Thompson Reuters Foundation): How 3,000 Companies Approach AI Governance

The Road to Accountable AI

Good data about how companies are implementing AI governance programs is essential both for organizations to benchmark their efforts, and for observers to understand the state of development. In this episode, Katie Fowler, Director of Responsible Business at the Thomson Reuters Foundation, joins Kevin Werbach to discuss the findings of Responsible AI in Practice, a new report drawing on a global dataset of roughly 3,000 companies across 13 sectors. Fowler unpacks the report's central finding: an enormous gap between corporate AI ambition and operational governance, with 44 percent of...

info_outline
Henry Ajder, Latent Space Advisory: Deepfakes and the Crisis of Digital Trust show art Henry Ajder, Latent Space Advisory: Deepfakes and the Crisis of Digital Trust

The Road to Accountable AI

AI-generated deepfakes are exploding in volume and quality, posing frightening challenges for public discourse, security, safety, and more. My guest, Henry Ajder, has been mapping the deepfake landscape since before most people had heard the term. In this conversation, he describes the dramatic changes in realism, efficiency, accessibility, and functionality of synthetic media tools since he published the first comprehensive census of deepfakes in 2019. Ajder describes the current moment as one of "epistemic nihilism," where people cannot reliably distinguish real from synthetic content and...

info_outline
Phil Dawson, Armilla AI: Insurance for AI Risks show art Phil Dawson, Armilla AI: Insurance for AI Risks

The Road to Accountable AI

Could a private insurance market play a significant role in compensating for AI-related harms and incentivizing companies to engage in more effective AI governance? Phil Dawson of Armillla AI explains why AI insurance is emerging as a distinct product category, why traditional policies aren't effective at addressing AI risks, and what AI insurance actually covers. Dawson details Armilla's journey from AI testing platform assurance provider to, managing general agent for AI insurance policies, arguing that the company's AI audit experience gave it the risk data and evaluation capabilities...

info_outline
Walter Haydock, StackAware: In Search Of AI Governance Certification show art Walter Haydock, StackAware: In Search Of AI Governance Certification

The Road to Accountable AI

Walter Haydock draws a direct line from military risk management to the enterprise AI challenge. His argues that organizations need to stop doing "math with colors," and move toward quantitative assessment that assigns dollar values to potential AI failures. Much of the conversation in this episode focuses on ISO 42001, the global standard for AI management systems, which Haydock has championed and which his own firm has gone through. He draws a three-part taxonomy of AI governance frameworks: legislation you either comply with or don't, voluntary self-attestable frameworks like the NIST AI...

info_outline
Richa Kaul, Complyance: Asking the Right Questions show art Richa Kaul, Complyance: Asking the Right Questions

The Road to Accountable AI

Richa Kaul breaks down the AI risk landscape for enterprises, and argues that the key to managing all of them is resisting the urge to sensationalize. Kaul offers a candid assessment of where enterprise AI governance committees are falling short, noting that many  lack the technical fluency to ask vendors the right questions, such as where customer data goes, whether it trains other clients' models, and what specific steps reduce hallucination. She suggests that market-driven security standards like SOC-2 and ISO 27001 often matter more in practice than government regulation, creating a...

info_outline
Michael Horowitz, UPenn: Governing AI That's Designed to Kill show art Michael Horowitz, UPenn: Governing AI That's Designed to Kill

The Road to Accountable AI

How AI is, could, and shouldn't be used in military and other national security contexts is a topic of growing importance. Recent conflicts on the battlefield, and between the U.S. military and a major AI lab, are forcing conversations about legal, ethical, and appropriate business limitations for increasingly powerful AI tools. Michael Horowitz, a Political Science professor and Director of Perry World House at the University of Pennsylvania, is one of the world's leading experts on military AI and autonomous weapons. In this episode, drawing on his two stints in the U.S. Department of...

info_outline
Tanvi Singh, Ekta AI: The Case for Sovereign AI show art Tanvi Singh, Ekta AI: The Case for Sovereign AI

The Road to Accountable AI

Tanvi Singh draws on over two decades of building and governing AI systems inside global banks to make a provocative case: you cannot be accountable for decisions you do not control. Enterprises are consuming intelligence through models they don't own, can't explain, and didn't train. Singh reframes sovereignty beyond data center locations and infrastructure, to control across the entire stack, so that an organization's AI reflects its own values, laws, and culture. Whlile frontier LLMs will continue to dominate the consumer and retail market, she argues that domain-specific models will...

info_outline
Ray Eitel-Porter, Co-Author of Governing the Machine: The Confidence to Use AI show art Ray Eitel-Porter, Co-Author of Governing the Machine: The Confidence to Use AI

The Road to Accountable AI

Ray Eitel-Porter, former Global Lead for Responsible AI at Accenture and co-author of the new book, Governing the Machine, discusses how enterprises can move from abstract AI principles to practical governance. He emphasizes that organizations can only realize AI’s benefits if responsibility is embedded into everyday business processes rather than treated as a standalone compliance exercise. Drawing on his experience leading global data and AI programs, Eitel-Porter explains how the release of ChatGPT transformed enterprise attitudes toward AI, accelerating adoption while exposing risks such...

info_outline
Alexandru Voica: Responsible AI Video show art Alexandru Voica: Responsible AI Video

The Road to Accountable AI

Alexandru Voica, Head of Corporate Affairs and Policy at Synthesia, discusses how the world's largest enterprise AI video platform has approached trust and safety from day one. He explains Synthesia's "three C's" framework—consent, control, and collaboration: never creating digital replicas without explicit permission, moderating every video before rendering, and engaging with policymakers to shape practical regulation. Voica acknowledges these safeguards have cost some business, but argues that for enterprise sales, trust is competitively essential. The company's content moderation has...

info_outline
Blake Hall: Safeguarding Identity in the AI Era show art Blake Hall: Safeguarding Identity in the AI Era

The Road to Accountable AI

In this episode, Blake Hall, CEO of ID.me, discusses the massive escalation in online fraud driven by generative AI, noting that attacks have evolved from "Nigerian prince" scams to sophisticated, scalable social engineering campaigns that threaten even the most digital-savvy users. He explains that traditional knowledge-based verification methods are now obsolete due to data breaches, shifting the security battleground to biometric and possession-based verification. Hall details how his company uses advanced techniques—like analyzing light refraction on skin versus screens—to detect...

info_outline
 
More Episodes

Kevin Werbach interviews Heather Domin, Global Head of the Office of Responsible AI and Governance at HCLTech. Domin reflects on her path into AI governance, including her pioneering work at IBM to establish foundational AI ethics practices. She discusses how the field has grown from a niche concern to a recognized profession, and the importance of building cross-functional teams that bring together technologists, lawyers, and compliance experts.

Domin emphasizes the advances in governance tools, bias testing, and automation that are helping developers and organizations keep pace with rapidly evolving AI systems. She describes her role at HCLTech, where client-facing projects across multiple industries and jurisdictions create unique governance challenges that require balancing company standards with client-specific risk frameworks. Domin notes that while most executives acknowledge the importance of responsible AI, few feel prepared to operationalize it. She emphasizes the growing demand for proof and accountability from regulators and courts, and finds the work exciting for its urgency and global impact. She also talks about the new chalenges of agentic AI, and the potential for "oversight agents" that use AI to govern AI. 

Heather Domin is Global Head of the Office of Responsible AI and Governance at HCLTech and co-chair of the IAPP AI Governance Professional Certification. A former leader of IBM’s AI ethics initiatives, she has helped shape global standards and practices in responsible AI. Named one of the Top 100 Brilliant Women in AI Ethics™ 2025, her work has been featured in Stanford executive education and outlets including CNBC, AI Today, Management Today, Computer Weekly, AI Journal, and the California Management Review.

Transcript

 AI Governance in the Agentic Era

Implementing Responsible AI in the Generative Age - Study Between HCL Tech and MIT