Terminator won’t save us
New technologies like AI open up a lot of chances but also bear many risks. Couldn’t we just learn how to deal with this from science-fiction films having already played through the major issues of AI? Unfortunately not. Rather than give advice, science-fiction distracts from the currently relevant ethical and legal challenges of AI and the socio-political implications.
Aya Jaff is a successful entrepreneur and coder working with machine learning – which is currently considered the most promising method to build artificial intelligence (AI). When she talks about her job, people would think at first that she creates Hollywood-like robots which might take over the world, as she recently told in a panel discussion. But what she actually does is writing computer programs to optimize very specific tasks.
Fear of takeover by machines is not a German phenomenon. However, Germans seem to be particularly afraid and skeptical of artificial intelligence when compared to other countries. These fears are echoed in the media when illustrating articles on AI with robots from the science-fiction (SF) genre – like the iconic Terminator, the innocent looking Ava from Ex Machina, the humanoid Cylons from Battlestar Galactica – or with similar looking ones. Rightly so, a group of German scientists who research into the social impact of robotics and AI established the Twitter hashtag #notmyrobots inviting everyone to post such unrealistic or misleading visualizations.
Science-fiction-style robots are not about tech, but about us
But why are science-fiction-style robots unrealistic or misleading? Isn’t the genre about future possibilities of scientific discoveries and technological progress? Yes and no. On the one hand, science-fiction surely builds on technology. It is exactly the current context of digitalisation, big data and machine learning which has led to an explosion of SF-films and series dealing with dystopian surveillance systems, robots and AI. From this vantage point SF refers to real world discussions on the possibilities but mostly the dangers of AI and can be interpreted as technology impact assessment.
But on the other hand, this view misses one central point: SF also uses technology as a means for conveying a certain narrative. The films and series are not about a realistic view of technology, they are about us. First of all, science-fiction-films are “disaster porn”. As early as 1965 Susan Sontag found that SF-films are not about science, but about the extensive disaster as form of art. But even more, the films and series are projections for primeval human desires and fears: we long to create an artificial human-like being and are worried at the same time of being dominated and destroyed by the forces we created. These motifs can be found throughout all kinds of different stories from Mary Shelley’s Frankenstein over E.T.A. Hoffmann’s The Sandman to the Jewish folklore of the Golem. Also, science-fiction critically reflects current social-political issues. In this sense especially the humanoid robot is a “narrative canvas” for the “other”, for discriminated and marginalized (human) beings; it confronts us with a critical humanism, in which our humanity defines itself within our attitude towards strangers and aliens.
We find all those motifs in the above mentioned examples: Terminator shows us in the light auf a disastrous apocalypse the human desire for a saviour; Ex Machina tells us a feminist inspired process of emancipation; Battlestar Galactica blurs the line between friend or foe where anyone could be an artificial cylon without even knowing it. Yet, in order to serve that kind of stories, AI in SF – be it a robot or not – must become almighty, magical or mystical.
Science-fiction distracts from the real challenges of AI
However, in real life AI is none of this. It is a technical tool already applied in many areas. We use AI in translation, in logistics, in detecting cancer, in recruitment and so on. AI has the potential to make our lives better. One of the downsides is a possible consolidation of discrimination through data bias, since machines are currently learning on the basis of data – which is nothing else than people’s digitalised experiences. This data carries the inequalities of history. If fed into a prediction system, this injustice will be transferred into the future. One of the examples is predictive policing in the US, where exactly those citizens come into the focus of the police who belong to a reference group that was supposedly criminal in the past – mostly Afro-Americans.
So, what we don’t need are anthropomorphized SF-robots with red eyes. What we need is a positive narrative for the future to use the opportunities of AI and to take measures to meet the challenges. This means diversity and a sense for the social context of data among the coders, an informed and attentive society and above all this means bold policy and smart regulation to make sure that our democratic values including the protection of minorities continue to count.
The many different commissions and strategies on AI on European and German level are not only indispensable to talk about the real – not the magical – challenges of AI. They are also crucial to start a broader discourse on AI and the economic, social as well as ethical implications. On this level, the German AI strategy – despite the fact that Germany could spend a lot more money – provides a promising starting point. People must not think of AI as almighty, mystical and uncontrollable force. This only distracts from the fact that technology is the product of our making with the aim to make our lives better.
 Irsigler, Ingo/Orth, Dominik (2018): Zwischen Menschwerdung und Weltherrschaft: Künstliche Intelligenz im Film, in: bpb (ed.): Aus Politik und Zeitgeschichte: Künstliche Intelligenz, 68. Jahrgang, 6–8/2018, p. 39–46.
 L. Meinecke and L. Voss (2018): ‘I Robot, You Unemployed’: Science-Fiction and Robotics in the Media. In J. Engelschalt et al. (Eds.), Schafft Wissen – Gemeinsames und geteiltes Wissen in Wissenschaft und Technik. Proceedings der 2. Tagung des Nachwuchsnetzwerks “INSIST”, München 2016, p. 208.
 Jackson, Patrick Thaddeus (2013): Critical humanism: theory, methodology, and Battlestar Galactica, in: Kiersey, Nicholas J./Neumann, Iver B. (eds.): Battlestar Galactica and International Relations, Routledge, New York, p. 18-36.
Dr. Isabella Hermann is Scientific Coordinator of the interdisciplinary research group “Verantwortung: Maschinelles Lernen und Künstliche Intelligenz” (Responsibility: Machine Learning and Artificial Intelligence) at the Berlin-Brandenburg Academy of Sciences and Humanities.
Title graphic: “Terminator – Arnold Schwarzenegger – anaglyph” by cobravictor on Flickr, Public Domain
This post represents the view of the author and does not necessarily represent the view of the institute itself. For more information about the topics of these articles and associated research projects, please contact email@example.com.
Sign up for HIIG's Monthly Digest
and receive our latest blog articles.
We approach the de-mystification of this claim by looking at concrete examples of how AI (re)produces inequalities and connect those to several aspects which help to illustrate socio-technical entanglements.
“System Risk Indication” (SyRI) deployed by the dutch government for automatically detecting social benefit fraud. The program was shut down due to a severe lack in transparency and unproportional collection...
AI won’t kill us in the form of a time-travelling humanoid robot with an Austrian accent. But: AI is used in various military applications – supporting new concepts of command…