The Clearsale Blog

How human biases can skew artificial intelligence tools

How human biases can skew artificial intelligence tools

Imagine a scenario where you need your car’s onboard navigation system to place an emergency call, but it won’t. Or arriving extra early for every international flight because airport security scanners never recognize your face.

For many people—especially people of color and women—these scenarios can be a frustrating reality. That’s because the AI that’s supposed to make life easier for us all doesn’t include diverse enough data to work for everyone. This is a big problem, but one that can be fixed.

Personal dignity, travel safety, and job hunting are just some of the aspects of living that can be improved with algorithms--if the technology learns to recognize and properly classify a full range of voice and faces. However, New York University’s AI Now center reported in April that a lack of diversity in the AI industry contributes to biased tools that inflict real-world harms. These biases can undermine the benefits that AI offers in so many areas of modern life.

How AI-powered biometrics can be biasedMultiple studies have found that facial and voice recognition algorithms tend to be more accurate for men than women. Facial recognition programs also have trouble correctly identifying transgender and nonbinary people. Image-recognition algorithms are often more accurate for people with lighter skin than people with darker skin. These discrepancies can create problems for users, ranging from inconvenient to potentially life-threatening.

For example, Georgia Institute of Technology researchers found that self-driving car safety technology doesn’t recognize pedestrians with dark skin as well as it spots white pedestrians. Why? The study authors say a teaching dataset with a lot of light-skinned people, coupled with too little emphasis on the darker-skinned people in the set, effectively teaches object-detection models to be better at recognizing white pedestrians. Self-driving car developers need to correct this disparity to prevent their cars from hitting people of color because they don’t recognize them as pedestrians and stop for them.

In airports, the Department of Homeland Security is testing facial-recognition biometrics to keep international air travelers safe. But this may result in more time-consuming and invasive screenings for flyers whose faces aren’t recognized properly by the AI. Some facial-recognition technology has trouble correctly identifying people of color and women—especially women with darker skin.

Like the biased pedestrian-recognition AI, these facial recognition algorithms were trained with datasets that skewed white and male. And as with pedestrian recognition, facial recognition algorithms need to learn from datasets that contain a fair mix of skin tones and genders.

Voice recognition technology is supposed to make lots of everyday tasks easier, like dictation, internet searches, and navigation while driving. However, since at least 2002, researchers and the media have documented cases of voice recognition working significantly worse for women than for men, because the algorithms are trained to recognize lower-pitched, typically masculine voices as human speech. The problem hasn’t yet been solved—a writer for the Guardian described in April the repeated difficulties her mother had telling her Volvo to make a phone call, until she deliberately lowered the pitch of her voice to sound more like a man.

University of Washington graduate student and Ada Lovelace Fellow Os Keyes has raised concerns that facial recognition systems could harm trans and nonbinary people in multiple ways. Security systems that scan faces to let residents into their apartment complex or monitor public restrooms could misidentify people who don’t clearly fit into one gender category. That, Keyes argues, could lead to more encounters with law enforcement, and raise the risk of arrest or injury.

It’s tempting to think that because algorithms are data-driven that they will generate impartial, fair results. But algorithms learn by identifying patterns in real-world datasets, and those datasets often contain patterns of bias—unconscious or otherwise. The challenge for AI developers is to find or build teaching datasets that don’t reinforce biases, so that the technology moves all of society forward.

Solving the AI bias problemUsing more inclusive datasets for AI learning can help create less biased tools. One startup, Atipica, is working to create an inclusive HR algorithm based on a broad dataset that pulls from resume and government data to accurately reflect workforce demographics.

Other steps are needed, too. Right now, the AI industry includes very few women, people of color, and LGBTQ people. A more diverse AI workforce would bring more perspectives and life experiences to AI projects. Transparency and bias-testing of AI systems can identify problems before products go to market and while they’re in use.

NYU’s AI Now center recommends bringing in experts in a range of fields beyond AI and data science to offer a broader perspective on how AI works in real-world situations. Some industry-watchers say regulation may be part of the solution. In Washington, some senators have proposed legislation that would require “algorithmic accountability” for companies that develop AI tools.

AI has the potential to make our lives safer and easier, if we train the systems to be inclusive and use them wisely. By taking steps now to eliminate biases in AI, we can make sure that advances in this technology move everyone forward. 

Original article at: https://www.information-management.com/opinion/how-human-biases-can-skew-artificial-intelligence-tools?brief=00000159-ffbf-d8bf-af7b-ffbf558d0000

You may also like

ClearSale Wins Comparably Awards for Best Work-Life Balance and Happiest Employees

ClearSale Wins Comparably Awards for Best Work-Life Balance and Happiest Employees

Fraud protection leader recognized by career site for a second time this year

3 Ways Tech Can Benefit Remote Teams

3 Ways Tech Can Benefit Remote Teams

Ecommerce businesses are used to an ever-evolving digital connection between them and their customers. But 2020’s COVID-19 pandemic has resulted in that digital connection making its way into the..

Shopping Habits by Gender: What’s Changed in 2020

Shopping Habits by Gender: What’s Changed in 2020

Do men hate shopping online? Are women more worried about fraud?

ClearSale Becomes Shopify Plus Certified App Program Partner

ClearSale Becomes Shopify Plus Certified App Program Partner

The fraud protection leader has been selected as a premier app provider for the highly regarded commerce platform.

Account takeover is the biggest fraud threat U.S. consumers haven't heard of

Account takeover is the biggest fraud threat U.S. consumers haven't heard of

Account takeover fraud is a huge problem, but most US consumers don’t know about it. Only 36% of US consumers say they are familiar with account hijacking fraud, even though it’s one of the..

Is Fraud Risk Scaring You Away From International Shipping?

Is Fraud Risk Scaring You Away From International Shipping?

With cross-border shopping estimated to make up 20% of e-commerce in 2022, many merchants are right to consider expanding into other countries. So what’s stopping them from pulling the trigger?

U.S. shoppers say they'll trade privacy (but not convenience) for better ecommerce fraud protection

U.S. shoppers say they'll trade privacy (but not convenience) for better ecommerce fraud protection

As more consumer spending shifts to e-commerce, merchants need to strike a balance between fraud protection and customer experience. A recent survey just before business closures swept the U.S...

What does effective B2C marketing look like now? Messaging is only part of the story.

What does effective B2C marketing look like now? Messaging is only part of the story.

As more consumers shop online, many companies are pivoting their marketing strategies to focus on digital channels. But smart marketing now requires more than simply reallocating resources for..

ClearSale and BigCommerce Partner to Prepare E-Commerce Merchants for the Holidays

ClearSale and BigCommerce Partner to Prepare E-Commerce Merchants for the Holidays

Fraud protection leader joins e-commerce platform powerhouse to help merchants accentuate customer experience while preventing fraud this seasonMIAMI, FL (September 11, 2020) -- Global fraud..

Want to write
for our blog?

Please review our writers' guidelines
https://www2.clear.sale/press/clearsale-guest-blog-guidelines
and then email guestwriter@clear.sale with your pitch!

Subscribe to our blog