Blog article
See all stories »

Is AI just as Biased as the Humans Creating It?

"The saddest aspect of life right now is that science gathers knowledge faster than society gathers wisdom."  Isaac Asimov

This is a reprint from article in ReadWrite

It might appear that data — the information you find in a scientific article, a history book, or a set of survey results — is just a collection of objective facts. The numbers, if sourced well, are supposed to be the hard truth untarnished by opinions, perspectives, or biases. In reality, this is rarely the case. Here is how AI can be just as biased as the humans creating it.

Data is only as good as the means used to acquire it, the methods employed to interpret and analyze it, and how it’s communicated. 

Missteps in these areas filter down into the conclusions we reach based on our data — and often result in the kinds of subjectivity and bias we want our “facts” to avoid.

Even with something simple like a survey, many factors — from the wording of the questions to the way we define our terms and the sample set we evaluate to form a judgment — potentially bring bias into our data. Bias impacts the conclusions we reach.

“I know there’s a proverb which says, ‘To err is human,’ but a human error is nothing to what a computer can do if it tries.” — Agatha Christie

Automated systems and artificial intelligence, which are devoid of human feelings, ideologies, and preferences, promise to help us avoid the subjectivity and bias inherent to human reasoning. However, humans must design and build these systems, which opens them up to the very same pitfalls.

The fact remains that the bias in AI has become such a widespread issue that Salesforce recently added an AI bias course to its educational platform, Trailhead, to teach entrepreneurs and other professionals about the dangers of homogenous data.

The Consequences of AI Bias

Biases in automated systems have already had adverse effects on certain groups. For instance, Wisconsin attempted to utilize AI in its court system to determine the likelihood a criminal would re-offend in order to help a judge determine sentencing or parole.

Unfortunately, the program was biased against people of color and incorrectly flagged black individuals as more likely to commit another offense than their white counterparts.

What about peer-review studies?

In another case, a peer-reviewed study determined that Amazon’s facial recognition technology had a harder time identifying women and individuals with darker skin. How did this happen? The training set — the example data fed into the system to teach the AI programming — lacked diversity.

Summary of court cases?

My company’s data extraction software is used for many purposes. One such goal is summarizing court cases for legal professionals. For the program to offer accurate summaries, the training set must include a broad range of case types. If the system were trained on only tort cases, for example, it would be far less effective in summarizing criminal cases.

Machine Learning AI — and other AI learning that’s hard to spot.

These examples show just how damaging AI systems based on machine learning can be; when the data used to set them up fails to adequately represent what they’re studying or is less diverse than the cases they attempt to evaluate, the results can be unfortunate. But AI bias isn’t always so easy to spot.

In business, for example, using a biased dataset could result in leadership greenlighting a project that’s doomed to fail. When the project inevitably flops, the execution team might take the blame when it should be placed on the faulty assumptions the plan relied on upon in the first place.

AI and automation can be hugely beneficial for businesses, and especially companies that rely on technology to perform work that used to require teams of people. 

However, if the technology isn’t’ constructed in a way that prevents bringing in bad data (or twists useful data), it might do more harm than good.

One proposed answer to this problem has been to encourage diversity across teams of data scientists — those who build the AI systems we rely on. The challenge is that we don’t always identify and promote the right kind of diversity to solve our bias problem.

Diversity Is More Than Skin Deep

When it comes to hiring data scientists, achieving what we usually think of as diversity (variance of ethnicity and gender) is not sufficient.

The value of diversity.

Of course, there’s value in having a workforce made up of men and women belonging to different ethnicities, religions, and cultures. However, diversity in these areas can still result in biased AI.

For example, a team of men and women from all around the world might build a system with gaps; they could all be trained using the same educational framework, for instance, or share the same personal views related to the items or issues in question.

People might share very similar approaches to solving a data problem if they’re products of the same culture, socioeconomic background, and political party. Homogeneity in these areas will inevitably bleed into the AI system the development team creates.

Avoid bias.

To avoid AI bias, we need to start thinking of diversity in terms of our ideas and points of view rather than just our skin tones and gender expressions. People who look different might very well think differently, too — but this isn’t always the case.

Achieving Diversity

It’s impossible for humans to avoid altogether subjective preferences, mistaken beliefs, and various holes in our thought processes. So it might be similarly unlikely to eliminate bias. But there are still steps we can take to mitigate the impact our preferences (think bias) have on the systems we create.

Diversity of thought is an excellent place to start. To put together data science teams with the right kind of diversity, hiring managers must focus on finding people with unique perspectives and unconventional ideas. 

These diverse attitudes, viewpoints, overviews, and objectivities often result from novel experiences and unorthodox training.

The challenge here, of course, is that a manager cannot ask candidates about their socioeconomic backgrounds, ages, cultural exposure, and religious and political beliefs during interviews.

How to diversify your team and all employees.

What they can do is ask more pointed questions related to the work potential hires would take on. For example, present candidates with an example of a data-modeling project. Ask them how they would go about framing the issue, gathering the data, and analyzing it.

If you already have team members who would approach the case using methods X and Y, consider it favorable if a candidate suggests a well-thought-out method Z.

An individual’s assertiveness is another key factor.

It does your team little good to add someone with a different way of thinking if she isn’t willing to speak up when she sees a problem. In order to avoid bias in analytical systems and AI tools, you need candidates who not only hold unique viewpoints, but also know how to express them.

Lastly, it’s important to work to cultivate a culture of openness and collaboration within your tech development and data science teams. When they can easily express viewpoints, they can discover and correct errors.

Remember: When a team shares an overarching goal of being as objective as possible, bias is minimized in the process.

AI and automation are on the upswing, and these systems promise to help minimize the impact of human subjectivity and bias while increasing our efficiency and productivity. To accomplish this goal, however, they need to be constructed with great care.

To move forward with care — we must take diversity very seriously. This means more than a simple focus on appearances when hiring data scientists and system architects. Leaders need to dig deeper, work to bring in different points of view and people who can express them, and support an inclusive development culture.

Teams with diverse ideas and a collaborative structure are far more likely to build AI systems that will truly transcend the biases that we humans struggle to eliminate in ourselves.

 

 

5868

Comments: (2)

Ketharaman Swaminathan
Ketharaman Swaminathan - GTM360 Marketing Solutions - Pune 11 March, 2020, 19:44Be the first to give this comment the thumbs up 0 likes

"People with middle initials are more credit worthy".

This is an insight uncovered by banks via Big Data analytics. As a result, if there's a toss up between making a loan to one person with middle initial and another person without middle initial, all other things being equal, a bank will make the loan to the former. To me, this is a best practice.

That said, since the decision is prejudiced against people without middle initial and is unfair to that cohort, it also qualifies as a bias according to the dictionary meaning of the term.  

Curious to know if there's an objective way to distinguish between Best Practice and Bias. Unless there is, we can keep talking about avoiding bias without really making any progress. End of the day, a whole lot of work in AI / ML is around analysis of data to uncover such insights.  

Sang Nkhwazi
Sang Nkhwazi - Sanrixa Ltd - Manchester 15 March, 2020, 08:27Be the first to give this comment the thumbs up 0 likes

Interesting. I think it also goes to the board level of a company. If the board of a company is not diverse, in terms of BOTH having enough women and people of colour at senior level, the advice that "flows" downstream may not be sufficient to empower managers to recruit people who will enable their AI / ML adoptions to be impartial or otherwise free of bias. It's not an exact science, but my point is the kind of leadership in a company will be crucial to ensuring that the AI / ML systems that are used, adopted or trained, do not end up exarcebating the exclusion and bigotry problems which society currently faces.

Now hiring