Responsible AI governance: The top 3 myths holding us back

There’s nothing more overhyped and less understood in the business world right now than artificial intelligence (AI). In a time when every new flashy startup claims to be an AI vendor, there are a ton of myths that need to be busted for us to apply AI where it matters most. Perhaps even more concerning: Many of these myths obscure what we should actually be learning about—responsible use of AI—and who should have a hand in building it.

As one of the first companies to bring AI to many of the world’s biggest brands, LivePerson (where I work as CMO) has been deeply involved in the fight against bias for years, including as a founding member of EqualAI. Since EqualAI’s launch five years ago, I’ve found that this nonprofit—which brings together experts across disciplines to reduce bias in the technology—always has new things to teach us about what’s real and what’s not in the AI space.

Recently, one of our LivePerson HR leaders, Catherine Goetz, completed EqualAI’s Badge Program, which educates leaders across wildly different functional areas and industries about how they can establish responsible AI governance frameworks at their own organizations. Catherine’s cohort included experts and executives from across telco, consumer packaged goods, defense, security, and tech companies, just to name a few. Together, they learned about some AI myths that all of us need to understand as we apply AI to our own spheres of influence:

Myth 1: AI isn’t my problem

Reality: AI is an “everyone” problem and an “everyone” opportunity. All companies should now consider themselves AI companies to some degree, because we should all be exploring and testing how it can help us do what we do best. But with AI more accessible than ever, there is no company that can avoid thinking deeply about the potential misuse of AI and its negative impacts.

Myth 2: Okay, so we need to do something, but our tech guys will handle it

Reality: We’re not going to code or program our way to responsible AI. Leaders across all functions (not just tech) need to play a part in establishing governance frameworks within their organizations. This obviously includes putting standards in place for product design, data integrity, and testing, among other things, but it also includes areas for teams like legal, HR, and recruitment to lead. Have you considered applicable laws around privacy? Do you have a designated point of contact for employees and customers? All functional areas have a role to play in making sure that you can safely stand by any AI you put out into the world.

Myth 3: We don’t need DEI to make AI

Reality: Diversity, equity, and inclusion (DEI) help make AI better. Full stop. One of the ways that we can be sure we don’t perpetuate historical and new forms of bias in AI is by making sure that the people developing these systems reflect the world at large, especially the populations that will use them to live, work, and play. Do you have the necessary diverse workforce to understand how your products and services impact different kinds of people that will—if you’re a successful business—use them every day?

Putting these myths to bed requires buy-in and action from cross-functional leaders at all levels of your business. That’s why several LivePerson leaders like Catherine are now badge-certified in responsible AI governance. They’ve learned about operationalizing AI principles, implementing tools to detect risks and biases, ensuring accountability, and creating a cohesive process to address potential harms. And their roles at our company are similarly wide-ranging, including leaders from HR, legal, product development, and engineering.

Today, there’s a serious lack of consensus when it comes to creating (let alone following) responsible AI standards, but leaders like Catherine are helping us make progress. Most recently, she coauthored a first-of-its-kind whitepaper from EqualAI called An Insider’s Guide to Designing and Operationalizing a Responsible AI Governance Framework. Working with cross-sectoral leaders in business (including Google DeepMind and Microsoft), government, and civil society, she helped develop a framework meant to apply to organizations of any size, industry, and maturity level. Their hope is that this framework can serve as a resource for any professional on the journey toward making the world better through more responsible AI.

I think this new whitepaper is also a powerful sledgehammer for busting persistent myths about AI in general, and who is responsible for making it responsible. AI can serve as a force for good in our world, and for our businesses, but there are profound implications if we fail to govern it effectively. Understanding that we’re all in this together will help usher us all into a safer and more responsible, AI-enabled future.

Ruth Zive is chief marketing officer at LivePerson and host of the Generation AI podcast.

https://www.fastcompany.com/90968258/responsible-ai-governance-the-top-3-myths-holding-us-back?partner=rss&utm_source=rss&utm_medium=feed&utm_campaign=rss+fastcompany&utm_content=rss

Établi 1y | 18 oct. 2023 à 12:40:04


Connectez-vous pour ajouter un commentaire

Autres messages de ce groupe

5 tips for mastering virtual communication

Andrew Brodsky is a management professor at McCombs School of Business at the University of Texas at Austin. He is also CEO of Ping Group and has received nume

23 févr. 2025 à 11:50:03 | Fast company - tech
Apple’s hidden white noise feature may be just the productivity boost you need

As I write this, the most pleasing sound is washing over me—gentle waves ebbing and flowing onto the shore. Sadly, I’m not actually on some magnificent tropical beach. Instead, the sounds of the s

22 févr. 2025 à 12:40:06 | Fast company - tech
The next wave of AI is here: Autonomous AI agents are amazing—and scary

The relentless hype around AI makes it difficult to separate the signal from the

22 févr. 2025 à 12:40:05 | Fast company - tech
This slick new service puts ChatGPT, Perplexity, and Wikipedia on the map

I don’t know about you, but I tend to think about my favorite tech tools as being split into two separate saucepans: the “classic” apps we’ve known and relied on for ages and then the newer “AI” a

22 févr. 2025 à 12:40:03 | Fast company - tech
The government or 4chan? The White House’s social media account is sparking outreach

The official White House social media account is under fire for posts that resemble something typically found on the internet forum 4chan.

A post shared on February 14, styled like a Val

21 févr. 2025 à 20:30:04 | Fast company - tech
How Wikipedia became a political lightening rod

Wikipedia has faced political threats for years, but this time, it may be at a breaking point.

Republicans have ramped up attacks against Wikipedia as yet another “

21 févr. 2025 à 18:10:17 | Fast company - tech