Making sense of our connected world
![Webseite – WHY_AI_Christoph Sorge Webseite – WHY_AI_Christoph Sorge](https://www.hiig.de/wp-content/uploads/bb-plugin/cache/Webseite-–-WHY_AI_Christoph-Sorge-panorama-5c7e6cf2f8960f356bc515006710f036-5ab2603d855d3.png)
Myth: AI Models are abstract and do not need personal data
In supervised machine learning, models are based on abstractions from training data. The models themselves, while structurally influenced by the training data, do not contain the data themselves. It therefore seems reasonable to treat data they contain as (almost) anonymous. However, this is not true. Research has shown that deanonymization is possible under certain circumstances. Therefore, the models have to be considered as partially containing personal data and data protection law has to be taken into account when developing AI models to safeguard data subjects.
Myth
AI Models are abstract and do not need personal data.
![](https://www.hiig.de/wp-content/uploads/2021/03/Busted-dark-tilted.png)
AI models are an abstraction which may or may not contain personal data. Data protection law needs to be taken into account.
Watch the talk
Material
Presentation Slides | |
KEY LITERATURE Shokri, R., Stronati, M., Song, C. & Shmatikov, V. (2016). Membership Inference Attacks Against Machine Learning Models. Al-Rubaie, M. & Chang, J. M. (2019). Privacy-Preserving Machine Learning: Threats and Solutions. EEE Security & Privacy, 17(2), 49-58. Liu, B., Ding, M., Shaham, S., Rahayu, W., Farokhi, F. & Lin, Z. (2021). When Machine Learning Meets Privacy: A Survey and Outlook. ACM Computing Surveys, 54(2), 1-36. |
About the author
![](https://www.hiig.de/wp-content/uploads/2021/08/Prof-Sorge_61437-1200x800.jpg)
Christoph Sorge
Professor, Saarland University (Chair of Legal Informatics), Saarbrücken, Germany
Christoph Sorge received his PhD in computer science from Karlsruhe Institute of Technology. He then joined the NEC Laboratories Europe, Network Research Division, as a research scientist. From 2010, Christoph was an assistant professor (“Juniorprofessor”) for Network Security at the University of Paderborn. He joined Saarland University in 2014, and is now a full professor of Legal Informatics at that university. While his primary affiliation is with the Faculty of Law, he is also a co-opted professor of computer science. He is an associated member of the CISPA – Helmholtz Center for Information Security, a senior fellow of the German Research Institute for Public Administration, and a board member of the German Association for Computing in the Judiciary. His research area is the intersection of computer science and law, with a focus on data protection.
This post represents the view of the author and does not necessarily represent the view of the institute itself. For more information about the topics of these articles and associated research projects, please contact info@hiig.de.
![HIIG-Newsletter-Header HIIG-Newsletter-Header](https://www.hiig.de/wp-content/uploads/2024/02/HIIG-Newsletter-Header.png)
You will receive our latest blog articles once a month in a newsletter.
Artificial intelligence and society
Navigating platform power: from European elections to the regulatory future
Looking back at the European elections in June 2024, this blog post takes stock of the Digital Services Act’s effect in terms of navigating platform power.
AI Under Supervision: Do We Need ‘Humans in the Loop’ in Automation Processes?
Automated decisions have advantages but are not always flawless. Some suggest a Human in the Loop as a solution. But does it guarantee better outcomes?
The plurality of digital B2B platforms
This blog post dives into the diversity of digital business-to-business platforms, categorising them by governance styles and strategic aims.