Skip to content
169 HD – Why AI
27 April 2021

Why, AI? – A new Online Learning Space

Unraveling Myths about Automation, Algorithms, Society and Ourselves


Everyone talks about AI. But how does it change society? And how can we use it to help society? For that, we need to understand how AI works and how we can make it work for us. We need to do it right because there is so much to get wrong. This is why we have put together the online learning space “Why, AI?”. It is conceptualized as an open education resource (OER) and all material is licensed under CC BY. Let’s continue a conversation about what we want AI to do – and what we want to do with AI. 

Artificial intelligence has become a major topic of both public debate and regulation. But what is “AI”? What are its limits? Do we need AI ethics? AI laws? On a global scale? In Europe? In Germany? Is AI neutral? Can AI discriminate? Can AI take decisions? Can AI technologies and applications  – do they have to be – explained or be explainable? What are the limits to the use of AI in different societal settings? Automated cars exist, but what about automated judges? 

Hosted by the Humboldt Institute for Internet and Society (HIIG) in the framework of the AI & Society Lab and the Research Group Global Constitutionalism and the Internet, “Why, AI?” is designed as a learning space to help you find out more about the myths and truths surrounding automation, algorithms, society and ourselves. 

Answering misconceptions with science – by means of an open course format

A stellar team of experts has contributed their time and intellectual energy to help deconstruct a great variety of urban legends about AI: that AI algorithms and automated decision-making systems are inherently bad, that they will end discrimination, that they will take over the world.
The truth is usually a bit less dramatic – but just as interesting. 

Explore all myths

“Why, AI?” is useful for journalists, educators, students, politicians and the broader public – to learn a bit more about how algorithms work and how AI can be shaped for future use, rather than allowing it to shape the future without our consent and control. 

Hans Jonas, a German philosopher of technology, reminded us to act in a way that the effects of our actions are compatible with the permanence of genuine human life. It is no exaggeration to say that we are at a key stage of technological development when it comes to AI. We as a society have to respond to progress in and with AI – but we need to do it right because there is a lot to get wrong. We hope that with “Why, AI?” you learn a little more about how AI impacts the world. And how we need to change AI before AI changes us.

“Why, AI?” is curated by Matthias C. Kettemann – head of the research group Global Constitutionalism and the Internet where he studies the rules of power and the power of rules in hyperconnected online spaces – and Daniela Dicks, Co-Lead of the AI & Society Lab, an interdisciplinary space exploring new perspectives on AI and mediating between different stakeholders in society that interact with AI.

Feedback? Ideas for a myth that should be busted? Get in touch at whyai@hiig.de.

 


The team behind this project: Frederik Efferenn, Lukas Fox, Christian Grauvogel, Katharina Mosene, Marie-Therese Sekwenz, Katrin Werner, Larissa Wunderlich.

It is hosted by the HIIG in cooperation with the Leibniz Institute for Media Research | Hans-Bredow-Institut (HBI), Junges Forum: Technikwissenschaften (JF:TEC), the Sustainable Computing Lab of the Vienna University of Economics and Business (SCL) and the Max Planck Institute for Comparative Public Law and International Law (MPIL) (Section International Law of the Internet)  & OESA e.V.


This post represents the view of the author and does not necessarily represent the view of the institute itself. For more information about the topics of these articles and associated research projects, please contact info@hiig.de.

Daniela Dicks

Fromer Co-Lead & spokesperson: AI & Society Lab

Matthias C. Kettemann, Prof. Dr. LL.M. (Harvard)

Head of Research Group and Associate Researcher: Global Constitutionalism and the Internet

Explore Research issue in focus

Du siehst Eisenbahnschienen. Die vielen verschiedenen Abzweigungen symbolisieren die Entscheidungsmöglichkeiten von Künstlicher Intelligenz in der Gesellschaft. Manche gehen nach oben, unten, rechts. Manche enden auch in Sackgassen. Englisch: You see railway tracks. The many different branches symbolise the decision-making possibilities of artificial intelligence and society. Some go up, down, to the right. Some also end in dead ends.

Artificial intelligence and society

The future of artificial Intelligence and society operates in diverse societal contexts. What can we learn from its political, social and cultural facets?

Sign up for HIIG's Monthly Digest

and receive our latest blog articles.

Further articles

two Quechuas, sitting on green grass and looking at their smartphones, symbolising What are the indigenous perspectives of digitalisation? Quechuas in Peru show openness, challenges, and requirements to grow their digital economies

Exploring digitalisation: Indigenous perspectives from Puno, Peru

What are the indigenous perspectives of digitalisation? Quechuas in Peru show openness, challenges, and requirements to grow their digital economies.

eine mehrfarbige Baumlandschaft von oben, die eine bunte digitale Publikationslandschaft symbolisiert

Diamond OA: For a colourful digital publishing landscape

The blog post raises awareness of new financial pitfalls in the Open Access transformation and proposes a collaborative funding structure for Diamond OA in Germany.

a pile of crumpled up newspapers symbolising the spread of disinformation online

Disinformation: Are we really overestimating ourselves?

How aware are we of the effects and the reach of disinformation online and does the public discourse provide a balanced picture?