Artificial intelligence or AI technology — what remained as a fantasy that existed just in books & movies is a reality today. Applications of AI have exponentially grown and have penetrated almost every industry. It’s just a matter of time for AI-enabled machines to replace humans in all the mundane jobs. Soon, AI bots would assist humans in executing tasks much efficiently.

As AI is transforming many sectors, and influencing our day-to-day life, one question raises.

Can we trust AI?

One of the many benefits offered by AI solutions is that it helps humans in decision-making. But, what if AI’s decisions are biased? Recent studies about algorithmic bias have pointed out such flaws in AI systems. These flaws termed algorithm bias could result in biased decision-making. But, how does it happen?

The root cause of bias in AI is human bias, which is an extensive issue and an ongoing research topic in psychology for years. A person can be biased and unaware of it. Unfortunately, the datasets, databases we have— reflect the bias of the person who curated it. When such datasets are used for training AI models, their decisions reflect the bias that was hiding in the dataset. Let’s see some examples;

In 2016 Microsoft launched an AI-powered conversational chatbot on Twitter that was designed to communicate with users via tweets and direct messages. However, within a few hours of its launch, it began responding with highly offensive and racist messages. The chatbot was trained using anonymous public data and had an internal learning function built-in. A group of people was able to manipulate the system with a concerted attack to inject racial prejudice into it. Some users were successful in flooding the bot with misogynistic, racist, and anti-semitic words.

In another example, Amazon worked for years on developing an AI model to automate the hiring process. However, they had to discontinue the automation tools as they exhibited bias. The AI-powered solution was supposed to sort through a stack of resumes and identify the best applicants. For that, the company fed the machine a decade’s worth of resumes from people who applied for various positions. But it was, later on, discovered that the algorithm restricted the applicants from women’s colleges as it was only favouring the words which were commonly found on men’s resumes.

Today, as more companies and organizations are planning to deploy AI systems across their operations, it is essential and an urgent priority for the stakeholders to be aware of the potential risks associated with it and how to mitigate them.

What is the source of bias?

•Underlying data with a limited number of variables: Models are trained on data that have a huge influence on societal or historical inequities and these datasets are limited to a short number of variables which results in ineffective results.

•Lack of diversity:  According to the Data from the Bureau of Labor Statistics it shows that the individuals who program are largely white men and only 12% of the leading machine learning programmers are women. This underrated problem would lead the results in favor of the majority and thus leading to an unreliable biasing

•Collection and selection of data from a particular domain set: Feeding the sample from a particular domain could increase the chance of biasing over a particular domain of preference. In the criminal justice sector, oversampling of data from certain neighbourhoods due to over-policing can result in recording more crimes.

In a nutshell, AI algorithms will learn the patterns of the data that is used to train it. In other words, AI models will have a bias that reflects the creators’ or data’s bias. Implying that cognitive prejudices are at the root of both modern AI and data biases.

Is it possible to create an unbiased AI?

Several approaches can be enforced to prevent the biasing constraints on AI models. The first approach consists of pre-processing the data to maintain quality accuracy while reducing the relationship between outcomes and sensitive characteristics. That is, to the best capability, remove the evident bias from the dataset.

The second approach consists of post-processing techniques in which models’ predictions are made to satisfy the fairness constraints. The fairness constraints would oversee the outcomes and can help to detect the bias. The third approach imposes fairness constraints on the optimization process to minimize the system’s ability to predict the sensitive characteristic.

Researchers and data scientists are working on designing and evaluating new approaches to achieve a more effective result. Here are some suggestive steps to reduce bias in AI

•Narrow down your business problem

Trying to solve too many scenarios also results in an unmanageable number of labels spread through an unmanageable number of groups. Identify challenges and model algorithms specific to the challenge.

•Curate data for a different opinion

For a single data point, there are usually several valid opinions or labels. Your model will be more flexible if you collect those opinions and account for legitimate, often subjective, disagreements.

•Recognize your training data

There are classes and labels in both academic and commercial datasets that can introduce bias into your algorithms. You’re less likely to be surprised by objectionable labels if you understand and own your data. Check to see if your data accurately reflects the diversity of your end-users.

•Consider all the end-users

Recognize that the end-users may not be identical to you or your team. Empathize with others. Recognize your end users’ diverse backgrounds, perspectives, and demographics. Avoid AI bias by anticipating how people that aren’t like you would communicate with your technology and the issues that might occur as a result.  

•Annotate with diversity

The more diverse your views are, the larger the pool of human annotators. This will significantly minimize bias, both during the initial launch and when the models are retrained. One choice is to use a global crowd of annotators, who can not only provide a variety of viewpoints, but also help a wide range of languages, dialects, and geographically specific languages.

•Test and deploy with feedback in mind

Models are seldom static throughout their lives. A popular, but serious, error is deploying the model without allowing end-users to provide input on how the model is performing in the real world. Opening a discussion and feedback platform will help to ensure that the model maintains optimum success levels for all.

•Make a plan to change the model based on the feedback

You should revisit the model regularly, not only based on customer reviews, but also by having independent people check it for updates, edge cases, instances of bias that you might have overlooked, and so on. Make sure you receive feedback from your model and provide feedback of your own to enhance its consistency, iterating toward greater accuracy at all times.

Is that it? No!

Businesses and policymakers can make an effort to subsidize these algorithm biases. They just have to look into things more carefully and here are some suggestions.

•Transparency about processes and metrics can help the team to understand the steps taken to promote fairness.

•Cognizant sampling and using third parties to audit data can improve data collection effectively.

•Be aware of the advancement of the fast-moving AI field and be cautious about the situation in which AI can rectify bias and in which AI can aggravate.

•Organizations need to stay up-to-date to see how AI can improve fairness.

•The collection of data should converge with more external variables and should be cumulated with more eligible data.

•Consider situations and use cases when an automated decision is made as promising systems use a combination of machines and humans to reduce bias.

•Invest more in biased research and research progress that will require interdisciplinary engagement with ethicists, social scientists, and experts.

To conclude, as we progress in identifying the bias points in AI, we should re-organize the standards to determine the fairness of human decisions and input as many datasets into the system as possible to reduce the biasing.

‍

Posted 
Feb 22, 2023
 in 
IT & Software
 category

More from 

IT & Software

 category

View All

Join Our Newsletter and Get the Latest
Posts to Your Inbox

No spam ever. Read our Privacy Policy
Thank you! Your submission has been received!
Oops! Something went wrong while submitting the form.