How to ensure fair AI throughout the supply chain

(Credit: Unsplash)

This article is brought to you thanks to the collaboration of The European Sting with the World Economic Forum.

Author: Mark Brayan, Chief Executive Officer, Appen

  • We need to go back to basics as we navigate the hype to make AI fair.
  • Basic fairness applies to all levels of the immense AI development lifecycle.
  • The fair treatment of the people who collect the data is being overlooked.

For some time now, there has been talk about how leaders developing AI applications need to build “fair AI”; this should be unbiased and equitable, ideally improving the quality of life of everyone it touches.

However, most of the thinking around ethical AI has been focused around models, explainability, technical teams and data governance.

But how about considering basic fairness in the AI development lifecycle? This lifecycle is immense. It entails the data collection and annotation contractors, the companies and individuals managing the data, the technology specialists building the AI models, the go-to-market experts building the AI applications, and the businesses and individuals using AI-powered products and services.

Such an approach is the only way we can ensure that technology continues to make the world a better place.

What is fair AI?

When an AI product is deployed in the real world, it must work as expected, deliver equitable results for all intended beneficiaries under all circumstances, and not harm anyone physically, mentally or emotionally. That’s a tall order. And it starts with building an unbiased and comprehensive data set.

While this may seem obvious, it’s easy for development teams to put unbiased data to the bottom of the pile and instead focus on achieving results as quickly as possible. However, doing so creates a long-term responsibility that developers can easily forget to maintain.

When is AI unfair?

Unfairness in the form of bias can appear in AI in multiple ways.

Consider criminal risk assessment algorithms that use behavioural and demographic data to determine the risk of reoffending.

A computer-based annotation system can’t do it alone when it comes to interpreting complex situations and catching subtle errors that could have real-life consequences. —Mark Brayan

One recent study found one such algorithm to be racially biased. For example, an 18-year-old black woman was charged with petty theft, having stolen an $80 bicycle. Despite having only one prior juvenile misdemeanour, she was rated as having a higher risk of reoffending than a 41-year-old white male who was charged with a similar crime but had several prior offenses.

Moving forward in time, the woman committed no other offenses while the male is now serving a prison term for the theft of thousands of dollars of electronics. Clearly, the algorithm was based on poor data.

Image: Proceedings of the National Academy of Sciences of the United States of America

Unfair AI also lives inside everyday technology. If a company sells cars equipped with speech recognition in multiple countries but trains the product using only native male speakers for each language, the system may struggle to understand women or anyone with a different accent.

This could lead to drivers being taken to the wrong destination. Worse still, it could cause distracted driving, leading to accidents. In addition to being unfair to some users, biased data can saddle solution providers with substandard products that can damage their reputation.

Where humans triumph

Developing a comprehensive and unbiased dataset requires data diversity and breadth. This ensures the product is trained in every situation it is likely to encounter in real life, such as all of the accents, voice tones and languages that a car’s speech recognition system may encounter in its target markets.

Achieving this means working with people who resemble the entire customer profile to collect, annotate and validate the AI model training data.

It also means working with a diverse team on the model building itself. A computer-based annotation system can’t do it alone when it comes to interpreting complex situations and catching subtle errors that could have real-life consequences.

For example, a human annotating images or a video for a self-driving car application could interpret that a person with a certain posture walking between two cars may be pushing a buggy that will appear in traffic before the person does.

Even the best computer-based annotation systems would struggle to make this interpretation. Similarly, a human reading a product review is much more likely to detect sarcasm than a machine is.

The people behind the data

Leaders committed to fair AI must include another important link in the AI development lifecycle when building global AI products or services: the millions of people who collect and label the data. Engaging these people in a fair and ethical way is mission-critical and should be part of every organisation’s responsibility charter.

Fair treatment means committing to fair pay, flexible working hours, including people from any and all backgrounds, respecting privacy and confidentiality, and working with people in a way that they feel heard and respected.

Leaders should also inspire their contractors in a way that instils pride in working on the most impactful technology used by the global economy.

Why does fair AI matter, beyond the obvious?

Quite simply, it’s good for society, and it’s good for business. Product teams, for example, are inspired when they’re building products that have a positive impact on their market and the world. But what else do fair products do?

  • They work for the entire target customer base: Products based on representative data will work for all users without bias, and so sell better, reduce frustration and lower returns.
  • They are safer: Comprehensive, unbiased training data will lead to safer, better-quality products, reducing the potential for failure.
  • They build loyalty: Great products and a great reputation are keys to increased customer loyalty.
  • They protect the brand: Products that work as expected often reduce the risk of serious and lasting brand damage.

According to one MIT Sloan study, only about one in ten enterprises currently report obtaining “significant” financial benefits from AI.

In 2021, as boards focus on closing the gap between AI’s potential and its reality, they will increasingly prioritise the adoption of the principles of fair AI. They know it will ensure projects work as designed, deliver expected benefits, and contribute to a better society.

Applications relying on AI are also infiltrating every industry, including the public sector. AI developers therefore have a certain responsibility to ensure their products are built on unbiased and comprehensive data sets that work for everyone.

Business and technology leaders should embrace fair AI as a core tenet to improve their businesses whilst helping society as a whole.

the sting Milestone

Featured Stings

Can we feed everyone without unleashing disaster? Read on

These campaigners want to give a quarter of the UK back to nature

How to build a more resilient and inclusive global system

Stopping antimicrobial resistance would cost just USD 2 per person a year

Ukrainian civil war: Is this the beginning of the end or the end of the beginning?

Syria: UN Humanitarian Coordinator calls for unimpeded access from within the country

Zimbabwe facing man-made starvation, says UN expert

Take-home pay growing at lowest level since 2008, as gender-gap persists: UN labour agency

European Youth, quo vadis?

Parliament votes reform for better European Co2 market but critics want it sooner than later

‘Pioneering’ former Chilean President Michelle Bachelet officially appointed new UN human rights chief

The EU will always have a stable partner in Montenegro, says President Đukanović

Statement by the European Commission following the extraordinary meeting of the EU-UK Joint Committee

Our healthcare systems are ailing. Here’s how to make them better

China has announced ambitious plans to cut single-use plastic

Canada grants asylum for Saudi teen who fled family: UNHCR

The experience economy is booming, but it must benefit everyone

Mobile technology: health in your hands

Draghi strives to control the unruly exploitation of financial markets by banking leviathans

Boardroom warriors: how CEOs are becoming champions of change

China dazzles the world with her Silk Road plan to connect, Asia, Europe and Africa

Where EU air pollution is deadliest

MEPs demand Bulgaria’s and Romania’s swift accession to Schengen area

European banking stress tests 2014: A more adverse approach for a shorter banking sector

5 lessons for social entrepreneurs on how to change the system

Hostilities in Syria’s southwest, mean cuts in vital aid across Jordanian border: Senior UN official

Human trafficking, slavery reports and health of migrants in Libya

The ECB must extend its money stimulus beyond 2018: Draghi reckoning

The role of junior entrepreneurs as a bridge between academia and business world

Air pollution: How to end the deaths of 7 million people per year

How the world can ‘reset’ itself after COVID-19 – according to these experts

Top envoy to Yemen praises ‘flexibility’ of chief negotiators as new UN mission chief is named

The priority for workplaces in the new normal? Wellbeing

Traffic congestion cost the US economy nearly $87 billion in 2018

A Valentine’s Special: giving back, a dialogue of love

Great Reset: Why LGBT+ inclusion is the secret to cities’ post-pandemic success

‘Dire consequences’ for a million children in the Middle East, North Africa, as funding dwindles

What is digital equality? An interview with Nanjira Sambuli

Millions denied citizenship due to ideas of national, ethnic or racial ‘purity’: UN rights expert

European Parliament backs CO2 emissions cuts for trucks

The hazards of “heroism” in the time of COVID-19

UN chief appoints Luis Alfonso de Alba as Special Envoy for the 2019 Climate Summit

Tsipras doesn’t seem to have learned his “almost Grexit” lesson and Greece faces again financial and political dead end

Financial services: Commission sets out its equivalence policy with non-EU countries

In The Bahamas, Guterres sees impact of ‘Category Hell’ hurricane, ‘powered by climate change’

Health Education, is it a necessity?

This farmer is saving the jungle by growing food in it

This company grows crops inside, stacked on top of one another

Finnish Council Presidency priorities debated in plenary

Anti-vaccination movement affecting youth in Europe

College meeting: European Commission reorganises the “Task Force for Relations with the United Kingdom” into the “Service for the EU-UK Agreements”.

Here’s what travelling could be like after COVID-19

Antarctica: the final coronavirus-free frontier. But will it stay that way?

Impressions of China

London wants new skyscrapers to protect cyclists from wind tunnels

A staggering one-in-three women, experience physical, sexual abuse

To keep track of the SDGs, we need a data revolution

Ebola situation worsening in DR Congo, amidst growing ‘funding gap’ UN health agency warns

EU Summit’s major takeaway: a handkerchief cannot save Greece from austerity

German heavy artillery against Brussels and Paris

ECB’s Draghi favours a cheaper euro to serve all Eurozone countries

The EU Parliament unanimously rejects Commission’s ideas about ‘seeds’

ECB intervenes to clean May’s and Schäuble’s mess

2030 development agenda: Major breakthrough for world of work

More Stings?


Speak your Mind Here

Fill in your details below or click an icon to log in: Logo

You are commenting using your account. Log Out /  Change )

Google photo

You are commenting using your Google account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s