By Janus Boye
In the past year Marli Mesibov has done significant research into chatbots, conversational interfaces, and voice UI. What she has found shows racism and bias in many areas of automated content. The reason is clear: an algorithm is only as strong as the strategy, taxonomy, and user testing that creates it.
In order to develop AI that is delightful, it must first be unbiased. We can do that in the people we hire to develop our products, and the ways in which we test and create the initial content.
In a recent member call, Marli reviewed concrete tactics for developing unbiased AI. Marli is Content Strategy Lead at Verily Life Sciences in Cambridge, MA and also a past Boye conference speaker.
The conversation started with voice user interfaces.
The problem with voice recognition
On her opening slide introducing herself, Marli actually put this as an extra bullet:
Super frustrated with my Google Assistant because it doesn’t listen to me
As she said, it’s often the case for women, that when the Google device doesn’t respond to them, a male voice will have a much higher success rate.
That’s because speech recognition, even here in 2022, performs worse for women than it does for men. There’s a bias in the AI that powers the content strategy, which is unfortunately optimised for that male version of the so-called accentless accent used by radio announcers and others considered flawlessly perfect, but unfortunately not the usual accent spoken by so many others.
Marli cited a Harvard Business Review article which covered this back in 2019. In the article titled Voice Recognition Still Has Significant Race and Gender Biases, the author opens with:
“As with facial recognition, web searches, and even soap dispensers, speech recognition is another form of AI that performs worse for women and non-white people.”
To illustrate the problem using numbers, Marli shared these 3 interesting figures for voice user interfaces:
95% overall success: Design in Tech report (PDF) from 2017 shows a 95% success rate in voice UI correctly responding to voices. In the five years since, no one has reported a higher success rate.
53% Scottish English. However, that rate goes down to 53% when we move from American English to Scottish English
13% lower for women. Most voice UI was tested with men, not women. Women’s voices have a 13% lower success rate.
Unfortunately, this is not specific to Voice UI! There’s bias across all channels.
Bias across all types of AI and UX
To get to the scale of the problem, Marli also shared how the content bias unfortunately exist across all types of AI and UX, including:
AI Assistants: AI assistants like Alexa and Siri need to respond to voice commands. If the AI assistant isn’t tested with a variety of accents and vocal pitches, it won’t work.
Chatbots: Chatbots are written with branching logic to respond to key words or phrases. The words and phrases are based on the creation team’s assumptions about what the end-user needs.
Data-based products: We are actively building products that take our assumptions, and build them into the branching logic and machine learning that we set up.
To illustrate the problem, Marli shared the example of Perspective, a free API that uses machine learning to identify "toxic" comments, making it easier to host better conversations online. To quote Marli:
APIs like Perspective initially failed, as they took broad social data, but didn’t comb it to remove the bias. The mistake: assuming data is neutral
The popular tweet from Jessamyn West back in 2017 shows how it went wrong.
Testing is key, but as Marli pointed out, the real problem is the lack of diversity in the teams building and testing the systems with:
9/10ths being cisgender men
9/10ths being able-bodied, white, American accents
These 9/10ths will have similar assumptions about phrasing and AI needs.
Finally, there’s also a common misunderstanding about the ‘edge cases’, with women being told to speak with a lower voice to get AI systems to hear them and apps being build for the 80%, with everything else as an “edge case”.
Let’s move onto Marli’s advice on what you can do.
What can you do?
Using the famous Gandhi quote - “Be the change you want to see in the world” - Marli shared good advice on what we can do in our own projects to make content strategy less biased.
First, she adviced to make the industry representative:
Hire diverse teams
Mentor people who don’t look or sound like you
Ask yourself why underrepresented groups are underrepresented
She also challenged us to look at our assumptions:
Create for “edge” cases
Focus on behaviors over demographics
Make your assumptions explicit, not implicit
And then, the final one and of my personal favorites: Keep learning
Read up on how to be anti-racist and anti-biased
Share your findings with coworkers
Avoid defensiveness in favor of being better
Learn more about bias in content strategy
During the call, Marli recommended Design for Real Life, a 2016 book by Eric A. Meyer and Sara Wachter-Boettcher which focuses on identifying stress cases and designing with compassion, so that you'll create experiences that support more of your users, more of the time.
Marli also shared this excellent guide with 5 Steps to Take as an Antiracist Data Scientist
Finally, you can also download the slides (PDF).