Digital age ‘desperately’ needs ethical and legal guidelines

Share

sophia_AI_robotcrp

This post Digital age ‘desperately’ needs ethical and legal guidelines was originally published on Horizon: the EU Research & Innovation magazine | European Commission.

Image credit – ITU Pictures, licensed under CC BY 2.0.

Jeroen van den Hoven, professor of ethics and technology at Delft University of Technology in the Netherlands, was speaking at a session on ethics in science and technology at the EuroScience Open Forum (ESOF) 2018, which is being held in Toulouse, France, from 9-14 July.

‘People are becoming aware that this digital age is not neutral…, it is presented to us mainly by big corporations who want to make some profit,’ he said.

He called for a Europe-wide network of institutions that can provide a set of values, based on the EU’s Charter of Fundamental Rights, which the technology industry could operate within.

‘We have to set up, as we’ve done for food, for aviation and for traffic, … an elaborate system of institutions that will look (at) this field of artificial intelligence.

‘We need to think about governance, inspection, monitoring, testing, certification, classification, standardisation, education, all of these things. They are not there.  We need to desperately, and very quickly, help ourselves to it.’

Prof. van den Hoven is a member of the European Group on Ethics in Science and New Technologies (EGE), an independent advisory body for the European Commission, which organised the session he was speaking at.

In March, the EGE published a statement on artificial intelligence (AI), robotics and autonomous systems, which criticised the current ‘patchwork of disparate initiatives’ in Europe that try to tackle the social, legal and ethical questions that AI has generated. In the statement, the EGE called for the establishment of a structured framework.

The European Commission announced on 14 June that they have tasked a high-level group of 52 people from academia, society and industry with the job of developing guidelines on the EU’s AI-related policy, including ethical issues such as fairness, safety, transparency and the upholding of fundamental rights.

The expert group, which includes representatives from industry leaders in AI such as Google, BMW and Santander, are due to present their guidelines to the European Commission at the beginning of 2019.

‘People are becoming aware that this digital age is not neutral…, it is presented to us mainly by big corporations who want to make some profit.’

Professor Jeroen van den Hoven, Delft University of Technology, Netherlands

Bias

Ethical issues surrounding AI ­– such as bias in machine learning algorithms and how to oversee the decision-making of autonomous machines – also attracted widespread discussion at the ESOF 2018 conference.

One major concern emerging with the fast-paced development of machine learning, is the question of how to account for the actions of a machine. This is a particular issue when using AI based on neural networks, a complex system set up to mimic the human brain that enables it to learn from large sets of data. This often results in algorithm becoming what is known as a ‘black box’, where it’s possible to see what goes in and what comes out, but not how the outcome was arrived at.

Maaike Harbers, a research professor at the Rotterdam University of Applied Sciences in the Netherlands, said that this was an important issue in the military, where weaponised drones are used to carry out actions.

‘In the military domain, a very important concept is meaningful human control,’ she said. ‘We can only control or direct autonomous machines if we understand what is going on.’

Prof. Harbers added that good design of the interface between humans and machines can help ensure humans exercise control at three important stages – data input, processing and reasoning, and the output or action.

Even in technologies that use AI for purposes that seem to be overwhelmingly positive, such as companion social robots for children, raise some tricky ethical issues. The conference audience heard that researchers working in this area are grappling with the effect these technologies can have on family relationships, for example, or whether they could create inequalities in society, or if they might create social isolation.

In the field of automated transport, researchers are also looking at the impact self-driving cars might have on wider issues such as justice and equality. They are investigating questions ranging from how to ensure equal access to new forms of transport to who should benefit from any cost-savings associated with automated transport.

However, the values we instil in AI may be a key factor in public acceptance of new technologies.

One of the most well-known moral dilemmas involving self-driving cars, for example, is the so-called trolley problem. This poses the question of whether an autonomous vehicle heading towards an accident involving a group people should avoid it by swerving onto a path that would hit just one person.

Dr Ebru Burcu Dogan from the Vedecom Institute in France, said research shows that while people were in favour of a utilitarian solution to the dilemma – for example, killing the driver rather than five pedestrians – they personally wouldn’t want to buy or ride in a vehicle that was programmed in such a way.

‘We all want to benefit from the implementation of a technology, but we don’t necessarily want to change our behaviour, or adopt a necessary behaviour to get there.’

If you liked this article, please share it.


This post Digital age ‘desperately’ needs ethical and legal guidelines was originally published on Horizon: the EU Research & Innovation magazine | European Commission.

 

Leave a Reply


Related Articles

2018 Ars Electronica Festival: Error – The Art of Imperfection
From Artificial Intelligence to Social Intelligence To err is human, it’s said. Could that be why we’re incessantly striving for perfection and steadfastly believe we can attain it with technology and science, and in spite of the fact that there is nothing that we fear more than being eliminated by a world of machinery that functions perfectly well without us? How can we rethink our very ambivalent relationship to technology as the driving force for configuring our future, and what errors ...
Still Life, Rhonda Holberton’s solo exhibition in NY
April 6 – June 2, 2018 Opening Reception, April 7 (check gallery website for hours) Transfer Gallery 1030 Metropolitan Ave  Brooklyn, NY 11211 CULT | Aimee Friberg Exhibitions is pleased to present Still Life, Rhonda Holberton’s third solo exhibition with the gallery, on view from April 6 – June 2, 2018 at Transfer Gallery, NY. Still Life features a networked video installation, prints rendered from augmented reality, immersive wallpaper constructed from bump map imaging, and gold mine...
Patchlab Digital Art Festival 2017 – Data art and artificial intelligence
Patchlab Digital Art Festival, organised in Krakow since 2012, is an international event dedicated to interdisciplinary (post)digital art forms, created at the intersection of art, new technologies and creative programming. During the 6th edition, taking place 24-29th October this year, the main focus will be at Data Art and AI. This time during the exhibitions there will be 14 installations focused on Data Art and AI including a unique installation 'Reading Plan' by Lien-Cheng Wang, showcas...
The National Gallery predicts the future with artificial intelligence
August 16 2017   The National Gallery, London, is working in collaboration with museum analytics firm, Dexibit, to use big data for predictive analytics. For decades, directors at the helms of the world’s cultural institutions have faced the challenge of balancing the historical and cultural objectives of telling curatorial stories with the economic needs of a museum dependent on a visiting public paying to visit temporary exhibitions and use its other commercial services. One of the most ...