A 5-step guide to scale responsible AI

(Credit: Unsplash)

This article is brought to you thanks to the collaboration of The European Sting with the World Economic Forum.

Author: Lofred Madzou, Project Lead, Artificial Intelligence and Machine Learning, World Economic Forum & Danny Lange, Vice-President, AI and Machine Learning, Unity Technologies

  • Deploying AI at scale will remain problematic until companies engage in a fundamental change to become ‘responsible AI’-driven organizations.
  • Companies should embrace this transformation as trust in AI systems will be the defining factor to determine who is worth doing business with.
  • Here is a guide to help them achieve responsible AI at scale.

Machine Learning is a revolutionary technology that has started to fundamentally disrupt the way that companies operate. Therefore, it is not surprising that businesses are rushing to implement it into their processes, as reported by the McKinsey & Company Global AI Survey. At the same time, a tiny percentage of these companies have managed to deploy Artificial Intelligence (AI) at scale – a process which seems harder to achieve given regular reports of unethical uses of AI and growing public concern about its potential adverse impacts.

These difficulties are likely to persist until companies engage in a fundamental change to become ‘responsible AI’-driven organizations. In practice, this requires addressing the governance challenges associated with AI, and then designing and executing a sound strategy. To help companies deploy responsible AI at scale, we offer a five-step guide.

AI creates unique governance challenges

We live in a world filled with uncertainty and the ability to build learning systems able to cope with this basic reality to a certain extent, by discovering patterns and relationships in data without being explicitly programmed, represents an immense opportunity.

However, there remain reasons for being concerned, because Machine Learning also creates unique governance challenges. For one thing, these systems are heavy-reliant on data, which incentivises companies to massively collect personal data, causing potential privacy issues in the process.

Second, collecting, cleaning and processing high-quality data is a costly and complex task. Consequently, business datasets often don’t accurately reflect the “real world”. Even when they do, they may simply replicate or exacerbate human bias and lead to discriminatory outcomes. That’s because the feedback loop in AI is likely to amplify any innate propensity embedded in the data.

Lastly, the power of massive computational systems with limitless storage capabilities eliminates the option of anonymity, as detailed personal behavioural information is taken into account to enable individual targeting at a previously unseen high level of granularity.

Risks that organizations consider relevant and are working to mitigate.
Risks that organizations consider relevant and are working to mitigate. Image: McKinsey & Company

More fundamentally, because AI-powered systems evolve with data and use, their behaviours are hard to anticipate; and when they misbehave, they are harder to debug and maintain. As opposed to classic software, one cannot simply correct the instructions given to the system to re-establish consistency with its intended functionality. Put simply, when something goes wrong, it’s harder to determine why it happened and implement corrective measures. In this context, an innocent objective such as maximising revenue could allow a highly capable AI learning system to develop deep and hard-to-detect ways to deceive the user into additional spending, which raises legitimate ethical concerns.

5-steps to deploy responsible AI at scale

As a company, how do you successfully deploy AI at scale while mitigating the risks discussed above? You should engage in a fundamental organisational change to become a responsible AI-driven company. To help navigate that change, we offer the following process as a starting point:

1. Define what responsible AI means for your company: To make sure the entire organization is pushing in the same direction, executives must define what constitutes responsible use of AI for their company through a collaborative process, involving board members, executives and senior managers across departments. This can take the form of a set of principles that guide the design and use of AI services or products. The drafting process of such principles should be structured around a practical reflection on how AI can create value for the organization and what risks (e.g. brand reputation, employee safety, unfair outcomes for customers, increased polarisation in the public discourse) need to be mitigated along the way.

Major industry actors, including Google and Microsoft, have already moved in this direction and released their responsible AI principles. More companies should follow their example. Drafting such principles provides two main benefits. First, it gives a chance to everyone, particularly top management, to get educated about responsible AI. Second, it could form the basis of a responsible AI business strategy, detailing how your organization plans to build a pipeline of responsible AI services and products.

2. Build organizational capabilities: Designing and deploying trustworthy AI systems should be an organization-wide effort. It requires sound planning, cross-functional and coordinated execution, employee training, and significant investment in resources to drive the adoption of responsible AI practices. To pilot these activities, companies should build an internal “Centre of AI Excellence”, which would concentrate its efforts on two core functions: training and driving adoption.

Indeed, to do their job, employees need to be trained to understand how risk manifests in their contextual interactions with AI systems and, more importantly, how to identify, report and mitigate them. That’s where even the most well-intentioned company can fall short if it focuses exclusively on technical teams. Also, the Centre should operate in close collaboration with business “champions” in charge of overseeing the implementation of trustworthy AI solutions and products.

3. Facilitate cross-functional collaboration: Risks are highly contextual, meaning diverse business functions have different risk perceptions. While designing your strategy, make sure to have complementary perspectives from various departments to develop a sound risk prioritisation scheme.

This will reduce top management “blind spots” and ensure stronger support from your workforce during the execution. Also, because learning systems tend to drive unanticipated behaviours, there will be risks that need to be addressed while the system is in operation. Here, close cross-functional collaboration, coordinated by risk and compliance officers, will be key for designing and implementing effective remedies.

4. Adopt more holistic performance metrics: Currently in the industry, AI systems are usually assessed based on their average performance on benchmark datasets. Yet AI practitioners and researchers acknowledge it is a rather narrow approach to performance assessment and are actively investigating alternative methods.

We suggest a more holistic approach: companies should, on a regular basis, monitor and assess the behaviour of their systems against their responsible AI principles. From that perspective, a system is deemed performant if its behaviour is consistent with the organizational definition of what is considered a responsible AI-powered service or product. Artificial Intelligence

What is the World Economic Forum doing about AI?

In 2019, the World Economic Forum’s Centre for the Fourth Industrial Revolution convened an informal multi-stakeholder group of leaders, known as the Global AI Council (GAIC) a keen interest in creating positive futures with advanced AI systems.

One of the goals of the Council is to provide strategic guidance to the global community on the priorities for AI governance and cooperation as well as the policy implications linked to advances in AI.

The project is taking place over several months and brings together a diverse group of individuals that includes science-fiction authors, economists, policymakers, and AI experts.

The council aims to open up the possibilities for its Positive AI Economic Futures using the creativity and expertise of these participants as well as opening up the process to a much wider range of contributors.

It is also in the process of initiating a second thread of the project, running in parallel with the workshops: a movie competition in partnership with the XPRIZE Foundation. Participants will create short movies showcasing their ideas for a future economy in a concrete form that speaks to individual aspirations and fears.

5. Define clear lines of accountability:Having the right training and resources is not enough to implement a lasting change if you fail to build the right lines of accountability. In other words, to do the right thing, employees must have the right incentives and be recognized for doing the right thing. Unsurprisingly, that’s one of the biggest challenges that Responsible AI practitioners are reporting. Here, we suggest two remedies.

First, you should introduce a vetting process, either as part of your AI products pre-launch review, or independent of it to make sure that ethical considerations have been addressed. This vetting process should be articulated with an organizational framework that maps the roles and responsibilities of each team involved and an escalation procedure to follow when/if there is persistent disagreement, for instance between product and privacy managers. Second, employees who have reported problematic use cases and took the time to introduce corrective measures should be rewarded as part of their annual performance assessment.

The way forward

There is an increasing awareness among business leaders that a responsible approach to AI is needed to ensure the beneficial and trustworthy use of this transformative technology. However, they are unsure about how to do this at scale while creating value for their companies. We want to reassure them that this is possible, but it requires profound organizational change.

As with any important change in life, the first steps are usually the hardest and we hope that our guide will help business leaders navigate that transition phase. We would also like to encourage them to persevere because, in the long run, responsible AI-driven companies are likely to be the most competitive. Indeed, the need for trust in AI systems is not a trend; it is the defining factor that will determine who is worth doing business with.

the sting Milestones

Featured Stings

Can we feed everyone without unleashing disaster? Read on

These campaigners want to give a quarter of the UK back to nature

How to build a more resilient and inclusive global system

Stopping antimicrobial resistance would cost just USD 2 per person a year

How COVID-19 is throttling vital migration flows

UN rights chief Bachelet appeals for dialogue in Sudan amid reports ‘70 killed’ in demonstrations

Asian and Pacific economies: decreases in tax revenue highlight need to broaden tax bases

Coronavirus: harmonised standards for medical devices to respond to urgent needs

Turkey: Extension of EU humanitarian programmes supporting 1.7 million refugees receives green light

Health services for Syrian women caught up in war, foster safety and hope: UNFPA

ILO: Progress on gender equality at work remains inadequate

EU plans to exploit the Mediterranean Sea and the wealth beneath it

An entrepreneurial point-of view on tackling the migration crisis and the risks of abolishing Schengen

So, what is your favourite Sustainable Development Goal?

Three ways to improve your corporate culture in the #MeToo era

Social inclusion: how much should young people hope from the EU? 

Athens urged to fast track asylum seekers amid island shelters crisis – UNHCR

Vaccination: understanding the challenges surrounding COVID-19 vaccination campaigns

What India’s route to universal health coverage can teach the world

To what extent are our moral standards responsible for killing people?

5 amazing schools that will make you wish you were young again

How to describe chronic pain beyond numbers? A Brazilian measuring instrument

Security Council must ‘come together’ to address the plight of children trapped in armed conflict, says UN envoy

Mental health as a tool of survival at the Pandemic

European Youth calls on European Council for urgent action on “humanitarian crisis” and questions the EU/Turkey deal respect of human rights

Future EU financing and recovery: MEPs to assess summit outcome

When will Eurozone’s unemployment rate stop being Europe’s worst nightmare?

Mobile World Congress 2015 first to debate EU’s new stance on Net Neutrality and Roaming Charges

Afghanistan: UN mission welcomes new polling dates following election delays

Fairness should be at the heart of the agricultural goods trade

Vaccination: understanding the challenges surrounding COVID-19 vaccination campaigns

A Sting Exclusive: “The Digital Economy and Industry are no longer opposing terms”, Commissioner Oettinger underlines live from European Business Summit 2015

Tiny Iceland teaches the West how to treat bankers

Women’s rights and how medical students can act as aides of progressive change

The results of Finland’s basic income experiment are in. Is it working?

We can decide to live within the limits of our planet

Promoting Health in the Brazilian Amazon: one nation but many cultures

ECB money bonanza not enough to revive euro area, Germany longs to rule with stagnation

New EU rules ensure better protection for 120 million holidaymakers this summer

EU Migrant Crisis: Italian Coast Guard Headquarters and Italian Navy to give host national opening addresses at Border Security 2016 in Rome

UN chief pays tribute to Egypt’s role in avoiding ‘dramatic’ escalation in conflict across the Gaza-Israel border

ACP-EU : Agreement on climate change, migration and post-Cotonou

CO2 can be a valuable raw material, not just a climate killer. Here’s how

Scientists have created biodegradable microneedles to fight eye disease

Annual UN women activists’ summit opens with focus on services, infrastructure

How start-ups will lead India through the Fourth Industrial Revolution

EU Ombudsman investigates the European Commission

EU and UK soon to be in a post-Brexit rush over free trade agreement with Australia

Commission and EIB provide CureVac with a €75 million financing for vaccine development and expansion of manufacturing

Could entrepreneurship be the real cure against the side effects of Brexit?

Rebuilding after COVID: The challenge is digital

Medical students: catalysts to close the gender gap

How the world can ‘reset’ itself after COVID-19 – according to these experts

Threat to biodiversity risks a flood of economic ruin

One third of poorer countries face both undernutrition and obesity: WHO report

State aid: Commission approves €900 million Slovenian scheme to support uncovered fixed costs of companies affected by coronavirus outbreak

Google case: A turning point in competition rules enforcement

Australia’s bushfires have pumped out half a year’s CO2 emissions

How Eurozone consumers spend their income when they have one…

25 years after population conference, women still face challenges to ‘well-being and human rights’, says UN chief

Cheese energy could power hundreds of UK homes

5G: How a ‘legion of robots’ could help save the rhino

The European Parliament fails to really restrict the rating agencies

Sponsored content: when QUALITY meets OPEX in manufacturing

More Stings?

Speak your Mind Here

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Google photo

You are commenting using your Google account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s