These rules could save humanity from the threat of rogue AI

robot 19

(Franck  V., Unsplash)

This article is brought to you thanks to the collaboration of The European Sting with the World Economic Forum.

Author: Johnny Wood, Writer, Formative Content


The possibility of man-made machines turning against their creators has become a trendy topic these days. Undoubtedly, Isaac Asimov’s Three Laws of Roboticsare no longer fit for purpose. For the sake of the global public good, we need something serious and more specific to safeguard our limitless ambitions – and humanity itself.

Today, the internet connects more than half the world’s population. And although the internet provides us with convenience and efficiency, it also brings threats. This is especially true in an age in which a good deal of our daily life is driven by big data and artificial intelligence. Algorithms have been widely used to determine what we read, where we go and how we get there, what music we listen to, and what we buy at what price. Self-driving cars, automatic cancer diagnosis and machine writing have never been so close to large-scale commercial application.

If data is the new oil, then AI is the new drill – and to extend this analogy, malfunctioning algorithms are the new pollution.

 Where is AI going to be most profitable?

Image: Statista / Tractica

It is important to note that malfunction does not equal malevolence. Likewise, good intentions do not guarantee a lack of legal, ethical and social troubles. Already with AI we have seen numerous examples of such issues, namely unintended behaviours, lack of foresight, difficulties in monitoring and supervision, distributed liability, privacy violation, algorithmic bias and abuse. Moreover, some researchers have started to worry about a potential rise in the unemployment rate caused by smart machines replacing human labour.

Troubles are looming

Misbehaving AI is increasingly prevalent these days. A facial recognition app tagged African-Americans as gorillas; another one identified 28 US Members of Congress as wanted criminals. A risk-assessment tool used by US courts was alleged as biased against African Americans; Uber’s self-driving car killed a pedestrian in ArizonaFacebook and other big companies were sued for discriminatory advertising practices; and lethal AI-powered weapons are in development.

We are marching into unmapped territory – which is why we urgently need rules and guiding principles as a compass to guide us in the right direction. Technology ethics are more important now than they have ever been, and must be at the core of this set of rules and principles.

We should acknowledge some of the early efforts to build such a framework. Notable examples include the Asilomar AI Principles, and IEEE’s ethics standards and certification programme.

And in late 2018, Pony Ma, the founder and CEO of Tencent, proposed an ethical framework for AI governance, namely ARCC (available, reliable, comprehensive and controllable).

Available, Reliable, Comprehensible, and Controllable: ARCC

Ma’s framework can become a foundation for the governance of AI systems in China and beyond. Its aim is to secure a friendly and healthy relationship between humanity and machinery in the thousands of years to come.

 AI should be available to all

Image: Tencent

Available.

AI should be available to the masses, not just the few. We are so used to the benefits of our smartphones and laptops, but more often than not we forget that half the world remains cut off from this digital revolution.

Advances in AI should fix this problem, not exacerbate it. We need to bring those living in developing areas, the elderly and the disabled into this digital world. The well-being of humanity as a whole should be the sole purpose of AI development. That is how we can ensure that AI will not advance the interests of some humans over the rest.

Take the recent development of medical bots as an example. Miying, Tencent’s AI-enabled medical diagnostic imaging solution, is currently working with radiologists in hundreds of local hospitals. This cancer pre-screening system has studied billions of medical images and detected hundreds of thousands of high-risk cases. The bot then refers these cases to experts. In doing so, it frees doctors from the daily labour of watching pictures and gives them more time to attend their patients.

Moreover, an available AI is a fair AI. A completely rational machine should be impartial and free of human weaknesses such as being emotional or prejudicial – but this should not be taken for granted. Recent incidents, like the vulgar language used by a Microsoft chatbot demonstrate that AI can go seriously wrong when fed by inaccurate, incomplete or biased data. An ethics by design approach is preferred here – that is, to carefully identify, solve and eliminate bias during the Ai development process.

Regulatory bodies are already formulating guidelines and principles addressing bias and discrimination. Firms such as Google and Microsoft have already set up their own internal ethical boards to guide their AI research.

 AI has to be safe and reliable

Image: Tencent

Reliable.

Since AI is already installed in millions of households, we need them to be safe, reliable, and capable of safeguarding against cyberattacks and other accidents.

Take autonomous driving cars as an example. Tencent is developing a Level 3 autonomous driving system and has obtained the license to test its self-driving cars on public roads in Shenzhen. But before getting the test license, its self-driving cars have been tested in a closed site for more than a thousand kilometres. Today, no real self-driving car is being commercially used on our roads, because the standards and regulations concerning its certification are still to be established.

Besides, for AI to be reliable, it should ensure digital, physical and political security, especially around privacy. We have seen cases where personal data is collected for training AI systems without the user’s consent. Therefore, AI should comply with privacy requirements, protect privacy by design, and safeguard against data abuse.

 AI should be understood better by all

Image: Tencent

Comprehensible

The enormous complexity of AI systems means this is easier said than done. The hidden layers between the input and output of a deep neutral network make it impenetrable, even for its developers. As a result, in case of a car accident involving an algorithm, it may take years for us to find the reason behind the malfunction.

Fortunately, the AI industry has already done some research on explainable AI models. Algorithmic transparency is one way to achieve comprehensible AI. While users may not care about the algorithms behind a product, regulators require deep knowledge of its technical details. Nonetheless, good practice would be to provide users with easy-to-understand information and explanations in respect of the decisions assisted or made by AI systems.

To develop a comprehensible AI, public engagement and the exercise of individuals’ rights should be guaranteed and encouraged. AI development should not be a secret undertaking by commercial companies. The public as end users may provide valuable feedback which is critical for the development of high-quality AI.

Tech companies should be required to provide their customers with information concerning the AI system’s purpose, function, limitations and impact.

 We need to be in charge - always

Image: Tencent

Controllable

The last – but not the least – principle is to make sure that we, human beings, are in-charge. Always.

Every innovation comes with risks. But we should not let worries about the extinction of humanity by some artificial general intelligence prevent us from pursuing a better future with new technologies. What we should do, is tmake sure that the benefits of AI substantially outweigh the potential risks. Top achieve that, we must establish appropriate precautionary measures to safeguard against foreseeable risks.

For now, people often trust strangers more than AI. We frequently hear that self-driving cars are unsafe, filters are unfair, recommendation algorithms restrict our choices, and pricing bots charge us more. This deeply embedded suspicion is rooted in information shortage, since most of us either don’t care or don’t have the necessary knowledge to understand an AI.

What should we do?

I would like to propose a spectrum of rules started from an ethical framework that may help AI developers and their products to earn the trust they deserve.

On the one side, we have light-touch rules, such as social conventions, moral rules and self-regulation. The ethical framework mentioned above fits here. At the international level, Google, Microsoft and other big companies have come up with their own AI principles, while Asilomar AI Principles and IEEE’s AI ethics programme are well praised.

As we move along the spectrum, there are mandatory and binding rules, such as standards and regulations. We wrote a policy report on self-driving cars this year and find that many countries are making laws to both encourage and regulate self-driving cars. In the future, there will be new laws for AI.

 

Further along the spectrum, there are criminal laws to punish bad actors for malicious use of AI. To the extreme right, there are international laws. For example, some international scholars have been pushing the United Nations to come up with a convention on lethal autonomous weapons, just like the Convention on Prohibitions or Restrictions on the Use of Certain Conventional Weapons.

Any new technology, whether a controlled nuclear reaction or a humanoid bot, is neither inherently good or bad. Ensuring it is the former is down to us.

the sting Milestone

Featured Stings

Can we feed everyone without unleashing disaster? Read on

These campaigners want to give a quarter of the UK back to nature

How to build a more resilient and inclusive global system

Stopping antimicrobial resistance would cost just USD 2 per person a year

UN must bring more women police officers into the fold to be effective – UN peacekeeping official

IMF: European banks do not perform their duty to real economy

A Europe that protects: Continued efforts needed on security priorities

Global warming: our responsibility

Why sustainable products are a win-win for all of us

Main results of Foreign Affairs EU Council, 16/07/2018

The COP24 Agreement: Yes, it happened at last

Afghanistan: UN mission condemns deadly attack near Kabul airport

International Women’s Day 2019: more equality, but change is too slow

Further reforms will move Slovakia toward a more innovative and inclusive society

The world is facing more disasters. This is how data can help us reduce that risk

Here’s how to check in on your AI system, as COVID-19 plays havoc

Big data is coming to agriculture. Farmers must set its course

Leaders need hard data to make the hard decisions about sustainability

The global response to the coronavirus pandemic must not be undermined by bribery

ECB guarantees the liquidity of the Atlantic financial volume

European Youth Capital 2019 announced: Novi Sad, Serbia

JADE Generations Club: Connecting perspectives, changing Europe.

Here are four steps SMEs can take for long-term success

Commission and EIB provide CureVac with a €75 million financing for vaccine development and expansion of manufacturing

What COVID-19 tells us about the changing nature of disaster risk

Greece and Ukraine main items on EU28 menu; the course is set

How privacy tech is redefining the data economy

Cyber attacks are shutting down countries, cities and companies. Here’s how to stop them

Will the end of QE come along with ECB’s inflation target?

World Malaria Day: 7 things to know about the deadly disease

A young person’s perspective on the Paris and Beirut attacks and aftermath

Convincing the Germans to pay also for the unification of Eurozone

A sterilised EMU may lead to a break up of Eurozone

US and Mexico child deportations drive extreme violence and trauma: UNICEF

Reintegrating former rebels into civilian life a ‘serious concern’ in Colombia: UN Mission chief

Combatting terrorism: EP special committee calls for closer EU cooperation

“If they think they can slave an entire nation, then they will just have the opposite results!”, Alexis Tsipras cries out from the Greek parliament

Hurricane Dorian: Bahamas death toll expected to rise as thousands remain missing

This is how travel hotspots are fighting back against overtourism

Berlin to pay at the end for Eurozone banks’ consolidation

‘Fire-fighting approach’ to humanitarian aid ‘not sustainable’: Deputy UN chief

The Working Methods of the von der Leyen Commission: Striving for more at home and in the world

Commissioner Hogan announces new transparency package

Syrian crisis is ‘clearest example’ of foreign investment in terrorism, Deputy Prime Minister says at UN

EU Commission: Germany can make Eurozone grow again just by helping itself

This robot has soft hands. It could be the future of sustainable production

Davos participants call for digital trade deal

4 key steps towards a circular economy

Tax revenues have reached a plateau

6 charts that show how Japan’s economy stacks up as it enters a new era

Migration and asylum: EU funds to promote integration and protect borders

Nicaragua: MEPs demand an end to repression of political opponents

The 10 most common types of plastic choking Europe’s rivers

In Pakistan, Guterres urges world to step up climate action, praises support to Afghan refugees

Protecting citizens’ access to social security in case of no-deal Brexit

China’s impact as a global investor; the Sting reports live from World Economic Forum 2015 in Davos

5 ways governments can unleash the power of young entrepreneurs

Europe must remember its past to build its future

The world wide web is 30 years old. What better time to fight for its future?

JADE Spring Meeting 2017– day 1: Excellence awards, panel discussion, keynote speeches

Population in crisis hit EU countries will suffer for decades

How a chocolate bar gives hope for a new economy

Five ways to increase trust in e-commerce

Fairer food supply chain: Agriculture MEPs clamp down on unfair trading

More Stings?

Advertising

Speak your Mind Here

Fill in your details below or click an icon to log in:

WordPress.com Logo

You are commenting using your WordPress.com account. Log Out /  Change )

Google photo

You are commenting using your Google account. Log Out /  Change )

Twitter picture

You are commenting using your Twitter account. Log Out /  Change )

Facebook photo

You are commenting using your Facebook account. Log Out /  Change )

Connecting to %s