Myth: AI can accurately predict and optimize human behavior
With technological advances in the field of AI and a growing amount of behavioral data that employees produce in their day-to-day routine, so-called people analytics tools have become a topic of public debate. These tools capture and analyze the behavioral data of employees, combine it with business data and offer employees and their manager’s insights into work routines, performance and potential. Based on this, people analytics promises to objectify and optimize employee-related decisions. Managers, therefore, place high expectations on these tools, especially with a growing number of employees who work from home and move outside their spatial control.
AI can accurately predict and optimize human behavior.
AI can indeed predict probabilities of human actions based on historical data. However, the accuracy of these predictions depends heavily on the data quality and is by no means error-free. As the behavior of humans in the workplace is complex and cannot always be quantified and measured, the current generation of AI can support people management in a limited area only and will certainly not make human managers obsolete.
Watch the talk
Giermindl, L. M., Strich, F., Christ, O., Leicht-Deobald, U. & Redzepi, A. (2021). The Dark Sides of People Analytics: Reviewing the Perils for Organisations and Employees. European Journal of Information Systems, ahead of print, 1-26.
Tursunbayeva, A., Di Lauro, S. & Pagliari, C. (2018). People analytics—A Scoping Review of Conceptual Boundaries and Value Propositions. International Journal of Information Management, 43, 224–247.
AlgorithmWatch. (2020). Positionen zum Einsatz von KI im Personalmanagement. Rechte und Autonomie von Beschäftigten stärken – Warum Gesetzgeber, Unternehmen und Betriebsräte handeln müssen.
BMAS. (2021). Gesetz zur Förderung der Betriebsratswahlen und der Betriebsratsarbeit in einer digitalen Arbeitswelt (Betriebsrätemodernisierungsgesetz).
Gal, U., Jensen, T. B. & Stein, M.-K. (2020). Breaking the vicious cycle of algorithmic management: A virtue ethics approach to people analytics. Information and Organization, 30(2), 1-15.
Hammermann, A. & Thiele, C. (2019). People Analytics: Evidenzbasiert Entscheidungsfindung im Personalmanagement, IW-Report, No. 35/2019. Köln: Institut der deutschen Wirtschaft (IW).
Leonardi, P. M. (2021). COVID‐19 and the New Technologies of Organizing: Digital Exhaust, Digital Footprints, and Artificial Intelligence in the Wake of Remote Work. Journal of Management Studies, 58(1), 249-253.
Netflix. (2020). Vorprogrammierte Diskriminierung [Video].
Nielsen, C. & McCullough, N. (2018). How People Analytics Can Help You Change Process, Culture, and Strategy. Harvard Business Review.
O’Neil, C. (2016). Weapons of Math Destruction: How Big Data Increases Inequality and Threatens Democracy. Crown Publishing Group.
Rasmussen, T. & Ulrich, D. (2015). Learning from Practice: How HR Analytics avoids being a Management Fad. Organizational Dynamics, 44(3), 236–242.
Stieler, W. (2021). Die Vermessung der Arbeit. MIT Technology Review, 4/2021, 22-27.
Thieltges, A. (2020). Machine Learning Anwendungen in der betrieblichen Praxis – Praktische Empfehlungen zur betrieblichen Mitbestimmung. Mitbestimmungspraxis, Nr. 33, Düsseldorf: Institut für Mitbestimmung und Unternehmensführung.
Tursunbayeva, A., Pagliari, C., Di Lauro, S. & Antonelli, G. (2021). The ethics of people analytics: risks, opportunities and recommendations, Personnel Review, ahead of print.
Zweig, K. (2019). Ein Algorithmus hat kein Taktgefühl: Wo künstliche Intelligenz sich irrt, warum uns das betrifft und was wir dagegen tun können. Heyne Verlag.
|UNICORN IN THE FIELD|
About the authors
Researcher, University of Hamburg / Associated Doctoral Researcher, Humboldt Institute for Internet & Society
Sonja is an associated doctoral researcher in the group Innovation, Entrepreneurship & Society at HIIG and researcher at the University of Hamburg. At HIIG, she currently supports the research project Artificial Intelligence & Knowledge Work, funded by the German Federal Ministry for Labour and Social Affairs. Her research focuses, among other things, on the digitalization of processes in human resource management and its impact on employees. In particular, she is interested in so-called people analytics applications. Prior to her role as a researcher, Sonja spent five years working as an HR practitioner in the field of HR information systems.
Researcher in the project “Anonymous Predictive People Analytics (AnyPPA)” at FZI Forschungszentrum Informatik, Berlin
Miriam is a research associate at the FZI Forschungszentrum Informatik Berlin/ Karlsruhe. She is currently supervising the BMBF-funded project Anonymous Predictive People Analytics. She focuses on the topic of the future of work and considers the social and ethical implications of the increasing use of AI in the workplace, as well as the development of co-determination and equal opportunities in the workplace.
This post represents the view of the author and does not necessarily represent the view of the institute itself. For more information about the topics of these articles and associated research projects, please contact email@example.com.
Sign up for HIIG's Monthly Digest
and receive our latest blog articles.
Whether civil society, politics or science – everyone seems to agree that the New Twenties will be characterised by digitalisation. But what about the tension of digital ethics? How do we create a digital transformation involving society as a whole, including people who either do not have the financial means or the necessary know-how to benefit from digitalisation? And what do these comprehensive changes in our actions mean for democracy? In this dossier we want to address these questions and offer food for thought on how we can use digitalisation for the common good.
Call for Proposals for a study on digital solutions and their impact on sustainability and rainforest protection in Indonesia.
Why is Artificial Intelligence so commonly depicted as a machine with a human brain? This article shows why one misleading metaphor became so prevalent.
Barriers in our physical environment are still widespread. While AI systems could eventually support detecting them, it first needs open training data. Here we provide a dataset for detecting steps...