Datafication and Consumer Trust
In einer datengetriebenen Wirtschaft und Gesellschaft befinden sich Verbraucher und Verbraucherinnen zunehmend in einer asymmetrischen Beziehung zu Unternehmen, die ihre Daten sammeln, meint Gastforscher Stefan Larsson. In einem Bericht für den schwedischen Verbraucherschutz hat er sich mit den Herausforderungen dieser Entwicklung beschäftigt.
We’re already living in a highly data driven society. Algorithms control the content flow displayed to us through search engines, for the news we read on Facebook and Twitter, the purchases we make via loyalty programs, the movies we watch on Netflix and the music we listen to on Spotify. Our behaviours are logged, collected and analysed by a growing number of operators for a growing number of reasons. At the same time, the data driven and automated society is also growing.
Seen from a consumer perspective, this naturally provides a number of services and innovations that we appreciate and often are prepared to pay for. But this data driven development also poses challenges to both consumer protection as well as how to develop the authorities’ supervision methods and role. Digital consumer profiling is in many senses the foundation of the digital economy, an economy that includes mega operators such as Google and Facebook, but also media houses that own a number of newspapers and media websites, relative newcomers in the sharing economy, the marketing industry, both e-commerce and traditional “brick-and-mortar” stores through, for example, loyalty cards and programs. In a new overview on digitalisation and consumer interest that I’ve done for the Swedish Consumer Agency, I highlight some developing trends that need to be looked into more closely. In short, how we manage consumer data and the operators and tools used for moderating, analysing and trading in the data will be of crucial importance for consumer status in the digital economy.
The need for critical studies
First of all, we can conclude that in committing to data driven innovation we also need to encourage a critical perspective and knowledge about how to manage data driven processes, algorithm controlled processes and data-analyses. We need to be able to recognise when the consumer needs protection and empowerment, evaluate winners and losers and strive for transparency both with regards to how the machines work as well as the regulations we would like to see guiding their work. This places demands on both legislators, politicians, supervisory authorities as well as on industry organizations and the academy. In brief:
- Seen from a consumer protection policy perspective, datafication entails a growing information asymmetry between the consumers and the market operators. We need to develop consumer protection, but the supervisory authorities also need to develop their supervisory and collaborative roles, both within and between authorities. This is as much an issue of power as of integrity and privacy.
- We, as consumers, commit to hundreds of agreements in the course of our daily digitalised lives. What are the implications of agreements that can only be understood and influenced by one of the parties? Seen from a consumer perspective, user agreements should certainly be shorter, clearer and easier to influence individually. But, are there any other ways of managing informed consent in a time of information overload? We need to figure this one out.
Consumer power also requires insight. For example, in 2015, the Norwegian Data Inspectorate (Datatilsynet) conducted a study of all the parties present when a user visited the index home page of six Norwegian newspapers. They concluded that between 100 and 200 cookies were downloaded to the user’s computer, that the user’s IP address was forwarded to 356 servers, and that on average 46 third parties were present at each of the automated ad trades taking place on the newspapers’ homepages. None of the six newspapers provided public information about the presence of such a large number of third-party companies. How can consumers be expected to choose safe services if they are unaware of the parties present, the information collected or what it’s used for?
Knowledge, transparency and balance
Addressing the challenge of consumers in a digital economy requires knowledge, insight and balance. We need to improve our knowledge about these relatively new developments and their implications – this means research, preferably within several disciplines that specialise in data, society, law, culture and economy. And these fields need to communicate much better than they generally do. The academic ways to organise, fund and publish tend not to help with regards to this. HIIG is a good example of such an interdisciplinary venture, but we need more. Much more. It should be the model, not the exception. At the same time, this poses a challenge to supervisory and organisational methods with regards to recognising downsides to automated processes. And finally, there is a constant need to maintain an articulated balance between the market and consumer protection – which is both a political and a legal matter. Amongst others, there is a risk that consumer trust in digital services will be weakened if the use of personal information is perceived as illegitimate. And as a result of weaker trust, the potential benefits of the digitally generated economy will also very likely be weakened.
- The Norwegian report The Great Data Race
- Larsson’s Swedish report on consumer interest and digitization, with an 8-page executive summary in English
Stefan Larsson is an Associate Professor in Technology and Social Change at Lund University Internet Institute (LUii) in Sweden, and a member of the Swedish Consumer Agency’s scientific council. From August to October 2016 he was a visiting researcher at the Alexander von Humboldt Institute for Internet and Society.
Photo: flickr.com CC BY-NC 2.0
Dieser Beitrag spiegelt die Meinung der Autorinnen und Autoren und weder notwendigerweise noch ausschließlich die Meinung des Institutes wider. Für mehr Informationen zu den Inhalten dieser Beiträge und den assoziierten Forschungsprojekten kontaktieren Sie bitte email@example.com
Bleiben Sie in Kontakt
und melden Sie sich für unseren monatlichen Newsletter mit den neusten Blogartikeln an.
JOURNALS DES HIIG
Seit Jahrzehnten wird gefordert, dass wir mehr Frauen in der Informatik brauchen, doch bis heute hat sich nichts getan. Deshalb kehren wir die Forderung um: wir brauchen weniger Männer in...
Wir entmystifizieren diese Behauptung, indem wir uns konkrete Beispiele dafür ansehen, wie KI Ungleichheiten (re)produziert, und verbinden diese mit verschiedenen Aspekten, die helfen, sozio-technische Verstrickungen zu veranschaulichen.
Die “System Risiko Anzeige” (SyRI) diente der niederländischen Regierung zur automatisierten Erfassung von Sozialhilfebetrug. Das Programm wurde allerdings wegen Intransparenz und unangemessener Datenerhebung verboten und beendet. Dies demonstriert, wie die...