Skip to content
169 HD – Why AI
27 April 2021

Why, AI? – A new Online Learning Space

Unraveling Myths about Automation, Algorithms, Society and Ourselves


Everyone talks about AI. But how does it change society? And how can we use it to help society? For that, we need to understand how AI works and how we can make it work for us. We need to do it right because there is so much to get wrong. This is why we have put together the online learning space “Why, AI?”. It is conceptualized as an open education resource (OER) and all material is licensed under CC BY. Let’s continue a conversation about what we want AI to do – and what we want to do with AI. 

Artificial intelligence has become a major topic of both public debate and regulation. But what is “AI”? What are its limits? Do we need AI ethics? AI laws? On a global scale? In Europe? In Germany? Is AI neutral? Can AI discriminate? Can AI take decisions? Can AI technologies and applications  – do they have to be – explained or be explainable? What are the limits to the use of AI in different societal settings? Automated cars exist, but what about automated judges? 

Hosted by the Humboldt Institute for Internet and Society (HIIG) in the framework of the AI & Society Lab and the Research Group Global Constitutionalism and the Internet, “Why, AI?” is designed as a learning space to help you find out more about the myths and truths surrounding automation, algorithms, society and ourselves. 

Answering misconceptions with science – by means of an open course format

A stellar team of experts has contributed their time and intellectual energy to help deconstruct a great variety of urban legends about AI: that AI algorithms and automated decision-making systems are inherently bad, that they will end discrimination, that they will take over the world.
The truth is usually a bit less dramatic – but just as interesting. 

Explore all myths

“Why, AI?” is useful for journalists, educators, students, politicians and the broader public – to learn a bit more about how algorithms work and how AI can be shaped for future use, rather than allowing it to shape the future without our consent and control. 

Hans Jonas, a German philosopher of technology, reminded us to act in a way that the effects of our actions are compatible with the permanence of genuine human life. It is no exaggeration to say that we are at a key stage of technological development when it comes to AI. We as a society have to respond to progress in and with AI – but we need to do it right because there is a lot to get wrong. We hope that with “Why, AI?” you learn a little more about how AI impacts the world. And how we need to change AI before AI changes us.

“Why, AI?” is curated by Matthias C. Kettemann – head of the research group Global Constitutionalism and the Internet where he studies the rules of power and the power of rules in hyperconnected online spaces – and Daniela Dicks, Co-Lead of the AI & Society Lab, an interdisciplinary space exploring new perspectives on AI and mediating between different stakeholders in society that interact with AI.

Feedback? Ideas for a myth that should be busted? Get in touch at whyai@hiig.de.

 


The team behind this project: Frederik Efferenn, Lukas Fox, Christian Grauvogel, Katharina Mosene, Marie-Therese Sekwenz, Katrin Werner, Larissa Wunderlich.

It is hosted by the HIIG in cooperation with the Leibniz Institute for Media Research | Hans-Bredow-Institut (HBI), Junges Forum: Technikwissenschaften (JF:TEC), the Sustainable Computing Lab of the Vienna University of Economics and Business (SCL) and the Max Planck Institute for Comparative Public Law and International Law (MPIL) (Section International Law of the Internet)  & OESA e.V.


This post represents the view of the author and does not necessarily represent the view of the institute itself. For more information about the topics of these articles and associated research projects, please contact info@hiig.de.

Daniela Dicks

Fromer Co-Lead & spokesperson: AI & Society Lab

Matthias C. Kettemann, Prof. Dr. LL.M. (Harvard)

Head of Research Group and Associate Researcher: Global Constitutionalism and the Internet

Sign up for HIIG's Monthly Digest

HIIG-Newsletter-Header

You will receive our latest blog articles once a month in a newsletter.

Explore Research issue in focus

Du siehst Eisenbahnschienen. Die vielen verschiedenen Abzweigungen symbolisieren die Entscheidungsmöglichkeiten von Künstlicher Intelligenz in der Gesellschaft. Manche gehen nach oben, unten, rechts. Manche enden auch in Sackgassen. Englisch: You see railway tracks. The many different branches symbolise the decision-making possibilities of artificial intelligence and society. Some go up, down, to the right. Some also end in dead ends.

Artificial intelligence and society

The future of artificial Intelligence and society operates in diverse societal contexts. What can we learn from its political, social and cultural facets?

Further articles

The picture shows a white wall with several clocks, all showing a different time. This symbolises the paradoxical impact of generative AI in the workplace on productivity.

Between time savings and additional effort: Generative AI in the workplace 

Generative AI in the workplace is enhancing productivity, yet employees face mixed results. This post examines chatbots' paradoxical impact on efficiency.

The picture shows seven yellow heads of lego figures, portraying different emotions. This symbolizes the emotions university educators go through in the process of resistance to change due to digitalisation.

Resistance to change: Challenges and opportunities in digital higher education

Resistance to change in higher education is inevitable. However, if properly understood, it can contribute to shaping digital transformation constructively.

The picture shows a young lion, symbolising our automated German text simplifier Simba, which was developed by our research group Public Interest AI.

From theory to practice and back again: A journey in Public Interest AI

This blog post reflects on our initial Public Interest AI principles, using our experiences from developing Simba, an open-source German text simplifier.