Business Model Approaches to Platform Regulation in the EU - Digital Services Act

Page created by Kurt Rose
 
CONTINUE READING
Business Model Approaches to
Platform Regulation in the EU
A framework for addressing disinformation and                                                 24 June 2020
hate speech on social media platforms
By John Albert, Adelaida Barrera, Charumati Haran,                                 #HertieDigitalGovernance
Maritta Rastuti, and two anonymous authors

    Looking ahead to the new Digital Services Act, one question facing
    policymakers is whether the underlying business models of social media
    platforms are misaligned with normative values of creating a safe, open, and
    reliable commons for online communication. This paper thus explores
    regulatory measures that would alter the business models of major social
    media platforms to protect users while continuing to promote innovation. It
    outlines three policy approaches for consideration by the European
    Commission, which are: reclassifying social media platforms as utilities;
    encouraging platforms to elevate authoritative journalism; and promoting
    alternative business models through a progressive tax on digital ad
    revenues. Adopting one or more of these approaches would serve to
    counteract harmful online content whilst promoting alternative models
    designed to serve the public interest.

Our vision
We aim to contribute to a respectful and reliable internet. Our mission is to share
information and knowledge about the development of the Digital Services Act and
thus raise awareness of how it relates to addressing online hate speech and
disinformation. Under the leadership of the German Presidency, the European
Commission will make crucial decisions that shape the development of a European
model for digital transformation. This emerging model is not only going to affect the
lives of people inside the European Union, it will also serve as a role model for other
countries in the World facing similar issues.

   We believe that more public discourse is needed to better inform the decisions
under deliberation by the European Commission. Toward this aim, we provide
information about ongoing developments and the different policy approaches that

24 June 2020                                                                                         1/11
Europe could take to tackle online hate speech and disinformation as part of the
Digital Services Act. We will contribute to this debate during the public consultation
phase of the European Commission on our website at www.digitalservicesact.eu.

  We are a group of graduate students at the Hertie School in Berlin, Germany, who
have teamed up with Daniela Stockmann, Professor of Digital Governance. The Hertie
School’s mission is to prepare students for leadership positions in government,
business and civil society institutions, to produce knowledge for good governance and
policymaking and to encourage responsible stewardship of the common good. We
consider ourselves as politically independent and are open to political positions from
the entire political spectrum in support of producing positive outcomes for society.

We are grateful for funding by the Hertie School’s Center for Digital Governance and
the Hertie School’s Student and Teaching Activity Fund. The Hertie School is a private
university based in Berlin, Germany, accredited by the state and the German Science
Council. The school was founded at the end of 2003 as a project of the Hertie
Foundation, which remains its major partner. We do not receive funding from
Facebook, Google, Twitter or any other tech company.

1 Introduction
Over the past 15 years, three companies - YouTube, Facebook, and Twitter - have
emerged as dominant platforms providing indispensable communications services to
the world. They also face the complex task of arbitrating online content.

Social media has demonstrably been exploited to disseminate false or hateful
information, which can promote conspiracy theories, influence election outcomes, or
incite real-world violence (Marwick & Lewis, 2017). This is exacerbated by online ‘echo
chambers,’ wherein users coalesce around like-minded groups that become both
targets and channels for spreading viral content (Schiffer, 2019). It has further been
shown that false news stories -- particularly those of a political nature that provoke
intense emotional reactions -- propagate faster and wider on social media than fact-
based stories (Vosoughi, 2018).

Such revelations have put pressure on social media companies to better moderate
their platforms in order to restore trust with users, advertisers, and governments.
Through self-regulation, these companies now regularly publish reports describing the
level of harmful content they remove from their pages, using trained human
moderators and algorithms based on machine-learning.

“In a lot of ways, Facebook is more like a government than a traditional company,”
Facebook CEO Mark Zuckerberg has said. Indeed, when major platforms set new
policies and modify their algorithms or product designs, they exercise tremendous
power through their massive user bases and impacts on adjacent businesses.

24 June 2020                                                                              2/11
However, such decisions are made by platform companies with little accountability to
users (Klonick, 2018, p. 1603). The reform of the Digital Services Act provides an
opportunity for Europe to adjust the rules governing companies to protect citizens’
fundamental rights and the public interest.

2 Business Model as Approach towards Platform
  Regulation

2.1     The current business model and targeted advertising
The main revenue stream for social media companies depends on their vast and
sophisticated processing of user data (Yakovleva & Irion, 2020). Platforms use data to
engage in ‘microtargeting’ - a lucrative means of personalising advertising content for
users. This model leads companies to optimize for user engagement (Zuckerman,
2019), and arguably infringes on the privacy and autonomy of users (Zuboff, 2018),
while potentially overexposing users to harmful content primed for virality on the
platforms (Silverman, 2016).

2.1.1    Effects on traditional news publishers
Quality journalism has an essential role to play in the fight against misinformation
(Crovitz, 2019). While social media platforms are arguably a huge benefit to news
publishers by sending them traffic (Jarvis, 2020), the paradigm shift towards targeted
online advertising has crippled the traditional advertising-driven business models of
established news publishers, with hosting, distribution, and monetization of news
content being ceded by publishers to platforms (Bell & Owen et al, 2017). This transfer
of power has contributed to waves of staffing cuts and closures of traditional news
outlets, particularly in local news (Hendrickson, 2019), which must now compete with
a proliferation of fake news content on social media that can sophisticatedly mimic
fact-based journalism (Recode, 2020).

2.2     Proposed alternatives to platform business models
In part to mitigate the spread of harmful online content, various politicians,
policymakers, and academics have suggested different ways to regulate platforms’
business models. U.S. Senator Elizabeth Warren has argued for breaking up platforms
to clear space for tech startups via antitrust measures. Margrethe Vestager of the
European Commission has discussed regulating companies’ revenue sources through
a subscription fee as a way to exempt users from tracking and advertising.

More far-reaching alternatives include a proposal by Tristan Harris for a new corporate
classification for platforms as “public utilities” which would subordinate companies’

24 June 2020                                                                              3/11
actions to the public or general interest. Landwehr, Borning, and Wulf (2019) present
alternative models of funding and control, such as publicly-funded platforms, NGOs
and cooperatives, and “No funding” (or minimally-funded) models such as Wikipedia.

3 Framework for Evaluation
The new Digital Services Act package aims to modernise the current legal framework
for digital services by placing two types of responsibilities on large online platforms: 1)
addressing the risks faced by their users and to protect their rights, and 2) ensuring fair
behaviour so that consumers have the widest choice and the Single Market remains
competitive and open to innovations (European Commission, 2020).

In line with the EU Commission’s pillars, we use the following evaluative framework
comprised of the three main principles guiding our policy recommendations:

3.1   Users’ right to privacy
We agree with the author Shoshana Zuboff that the issue of individual privacy is a key
problem in social media platforms’ current business models. We believe that
companies should respect and protect users’ right to privacy. Protection for privacy is
not only an end in itself (Article 8 of ECHR) but also essential to reclaim users’
autonomy and self-determination that has been lost due to the constant push of
information from the platforms (Napoli, 2019).

3.2   Protecting the public interest
As Daniela Stockmann (2020) argues, platforms’ double roles as data companies and
information services have led to a tension between their private and public interests.
The imperative to safeguard the public interest can justify the need for government
intervention in the market (Nooren et al, 2018). The use of social media platforms
further relates to the public interests of freedom from improper influence (Helberger,
Kleinen-von Königslöw, & van der Noll, 2015) in the use of data, as well as to pillars of
the EU community such as free democratic elections and access to reliable
information.

3.3   Ensuring innovation and sustainable business practices
As the EU Commission has stressed (COM(2016)288), the aim of strengthening the
single market in relation to online platforms should be guided by the goal of limiting
the concentration of market power by large companies, without overburdening or
disincentivizing new innovators from entering the market (Dittrich 2018). While
reducing profit margins or market power for major platforms should not be the goal of
regulation, we believe this should be an acceptable outcome if it ensures users’ rights
to privacy and protects the public interest.

24 June 2020                                                                                  4/11
4 Policy Suggestions

4.1     Reclassifying social media companies as utilities
This approach would reclassify major platforms as distinct economic entities bound by
specific standards, obligations, and oversight as public service providers (Sabeel,
2018). This would enable regulators to comprehensively evaluate and regulate social
media platforms’ business practices, including how they respond to disinformation
and hate speech.

Building on digital rights-based approaches (such as access to the internet and
freedom of expression), the activities of social media companies can be considered an
essential public service. Recognizing this, some critics have already argued for a
reclassification of large platform companies as utilities. This includes proposals to
break up companies through antitrust proceedings; however, antitrust measures
arguably lead to a diffusion of the problems of harmful content across a multitude of
firms, making it more difficult to track and address (Landwehr, Borning & Wulf, 2019).
As utilities, firms could instead be broken up by functional capabilities, restricting the
data that the common carriers and their analogs can share (ibid).

4.1.1    Social Impact Assessments
Another requirement of a utility-model approach could be to subject the platforms to
social impact assessments (Harris, 2020), in which new products would be evaluated
and approved based on their potential impacts on dimensions of individual and
societal health, including the impacts of disinformation and hate speech.

4.2     Leveling the playing field for news publishers
This approach addresses the ways in which social media platforms could appropriately
compensate traditional media companies from advertising revenues generated when
users search for and consume news articles and videos via the platforms. This would
follow Australia’s 2020 order to draft a “Mandatory Code of Conduct'' for platforms on
issues including the sharing of data with traditional publishers, ranking of news
content online and the sharing of revenue generated from news. A similar measure
taken by the French Competition Authority in April enforces a 2019 European
Commission change to copyright law. As with the Australian model, this measure
could be enforced through penalties and sanctions that include a dispute resolution
process.

We believe this kind of policy intervention would encourage platforms to strengthen
authoritative journalism by pushing them to coordinate with news publishers that are
subject to stricter standards, are liable for their content, and serve as a bulwark against
misinformation.

24 June 2020                                                                                  5/11
4.3     Supporting alternative business models through a progressive tax on digital
        ad revenues
As proposed by Nobel Prize-winning economist Paul Romer (2019), a progressive tax
on targeted ad revenues could accomplish ends similar to those sought by antitrust
measures without necessitating a judicial process. The Romer model suggests that
higher tax rates for larger companies would a) discourage large mergers and
acquisitions and b) make it easier for new companies to enter the market, encouraging
greater competition that would ultimately serve consumers. This measure could
incentivize companies to adopt business models that do not rely on targeted
advertising, and tax revenues could be reinvested to fund new “public service media”
models (Zuckerman, 2020) as well as comprehensive study of the social and individual
effects of digital media.

4.3.1    Incentivizing platform companies to adopt subscription models
In order to avoid a progressive tax on digital ad revenues, platforms could be
incentivized to adopt more traditional practices like a subscription model (Romer,
2019). A subscription fee would alter the business model of platforms by reducing or
removing their reliance on targeted advertising. Limiting data collection would
protect users from targeted ads that infringe on their privacy and autonomy (Opinion
3/2018 EDPS Opinion on online manipulation and personal data, 2018). This could
disrupt the ability of bad-faith actors to target susceptible audiences through
disinformation, and might shift platforms’ incentives away from engagement-centric
metrics that reward harmful content.

4.3.2    Reinvesting tax revenues in new business models and research
As proposed by Ethan Zuckerman (2020), one possible use of digital ad tax revenues
could be to facilitate experimentation and the development of new forms of public
service digital media. In order to guide this proposed experimentation, he argues,
governments should also fund extensive research on the effects of social media on our
individual and societal health. In this way, new kinds of public platforms could be
created as an alternative to surveillance-based business models, based on sound
scientific evidence that addresses the root causes of problems such as the proliferation
of harmful content.

4.3.3    Adopting a regulatory sandbox approach to support new ventures
To realise the social and economic potential of the preceding policy suggestions, there
is a need for an agile regulatory framework to keep pace with emerging technological
innovations while mitigating their harmful impacts. One model for such a framework
comes from finance, as lawmakers across the world have already started applying
‘regulatory sandboxes’ to regulate FinTech companies.

24 June 2020                                                                               6/11
Using this framework, policymakers in the EU can experiment with eliminating red
tape during a limited time (Federal Ministry for Economic Affairs and Energy, 2019),
facilitating development of new tools (Allen, 2015), testing safeguards (Monetary
Authority of Singapore, 2020) and enabling real-time proactive policy change (Arner
et al,2016).

With the launch of the European Forum for Innovation Facilitators (“EFIF”) in 2019 to
foster collaboration between European supervisors and FinTech firms, there is already
a forum that can be modified and expanded to support public-private partnerships
around social media platforms operating in the public interest.

5 Conclusion
The introduction of the Digital Services Act is an opportunity to update the rules
governing social media platforms by encouraging new approaches to regulation that
better protect the public interest while promoting innovation under the single market.
Our accordant policy recommendations have ranged in severity, from those that
would reform current practices (such as reclassifying platform companies as utilities
or requiring them to more generously compensate news publishers) to developing
alternative business models through a progressive tax on digital ad revenues.

There is no doubt that today’s major social media platforms have generated benefits
for their users, opening up entirely new forms of communication. However, we see
that the harms that emerge from hate speech, disinformation and misinformation on
the platforms risk may be too complex for companies alone to manage, especially
given their existing business models that may be misaligned with the public interest
around these issues.

Any new regulatory framework should take account of such complexity and invest in
more research to better understand the root problems, foster targeted innovations,
and provide the public with greater choices with which to navigate the informational
commons safely and reliably.

6 References
Allen, H.J. (2015) ‘A US Regulatory Sandbox?’, SSRN Electronic Journal. [online] Available at:
https://www.fca.org.uk/publication/research/regulatory-sandbox.pdf. (Accessed: 8 June
2020)

Arner, D.W., Barberis, Janos Nathan and Buckley, R.P. (2016). FinTech, RegTech and the
Reconceptualization of Financial Regulation. [online] Ssrn.com. Available at:
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=2847806 (Accessed 8 June 2020)

24 June 2020                                                                                     7/11
Bell, E., & Owen, T. et al (2017). ‘The Platform Press: How Silicon Valley Reengineered
Journalism’. Tow Center for Digital Journalism. [online] Available at:
https://www.cjr.org/tow_center_reports/platform-press-how-silicon-valley-reengineered-
journalism.php, (Accessed: 25 May 2020).

Christl, W. (2017). ‘How companies use personal data against people’, Cracked Labs - Institute
for Critical Digital Culture. [online] Available at:
https://crackedlabs.org/dl/CrackedLabs_Christl_DataAgainstPeople.pdf (Accessed: 7 Jun.
2020).

Crovitz, L. (2019). ‘Fighting misinformation requires journalism, not secret algorithms’
NiemanLab. [online] Available at: https://www.niemanlab.org/2019/12/fighting-
misinformation-requires-journalism-not-secret-algorithms/. (Accessed: 5 June 2020).

Council of Europe, European Convention for the Protection of Human Rights and
Fundamental Freedoms, as amended by Protocols Nos. 11 and 14, 4 November 1950, ETS 5.
Available at: https://www.refworld.org/docid/3ae6b3b04.html (Accessed: 25 May 2020).

Dittrich, P. (2018). ‘Online Platforms and how to regulate them: an EU Overview’, Jacques
Delors Institute, Policy Paper No. 227, June 14, 2018.

European Commission (2016). ‘Communication from the Commission to the European
Parliament, the Council, the European Economic and Social Committee and the Committee
of the Region. Online Platforms and the Digital Single MarketOpportunities and Challenges
for Europe (COM[2016] 288 final)’ [online] Available at:
https://ec.europa.eu/transparency/regdoc/rep/1/2016/EN/1-2016-288-EN-F1-1.PDF
(Accessed: 25 May 2020).

European Commission. (2016). Code of Practice on Disinformation. [online] Available at:
https://ec.europa.eu/digital-single-market/en/news/code-practice-disinformation (Accessed:
25 May 2020.)

European Commission. (2020). The Digital Services Act package. [online] Available at:
https://ec.europa.eu/digital-single-market/en/digital-services-act-package [Accessed 8 Jun.
2020].

Financial Conduct Authority UK. (2017). Sandbox tools. [online] Available at:
https://www.fca.org.uk/firms/innovation/regulatory-sandbox-tools (Accessed 8 June 2020).

Harris, T. (2020) ‘EU Should Regulate Facebook and Google as 'Attention Utilities'’, Financial
Times, 1 Mar. [online] Available at: https://www.ft.com/content/abd80d98-595e-11ea-abe5-
8e03987b7b20 (Accessed: 15 March 2020).

Haydar, N. (2020) ‘Facebook and Google to Face Mandatory Code of Conduct to 'Level
Playing Field' with Traditional News Media’, ABC News, 20 Apr. [online] Available at:
www.abc.net.au/news/2020-04-20/facebook-and-google-to-face-mandatory-code-of-
conduct/12163300 (Accessed: 4 May 2020)

Helberger, N., Kleinen-von Königslöw, K., & van der Noll, R. (2015) ‘Regulating the new
information intermediaries as gatekeepers of information diversity’, The journal of policy,

24 June 2020                                                                                     8/11
regulation and strategy for telecommunications, information and media. [online] Available
at: https://doi.org/10.1108/info-05-2015-0034, (Accessed: 25 May 2020)

Hendrickson, C. (2019) ‘Local journalism in crisis: Why America must revive its local
newsrooms’, The Brookings Institute, 12 Nov. [online] Available at:
https://www.brookings.edu/research/local-journalism-in-crisis-why-america-must-revive-its-
local-newsrooms/ (Accessed: 8 June 2020)

Jarvis, J. (2020) ‘Jeff Jarvis talks about why forcing the platforms to fund the media is wrong’,
Galley By CJR, 20 May. [online] Available at: https://galley.cjr.org/public/conversations/-
M7lwbaLbW2qgnDinuIu (Accessed: 6 Jun. 2020).

Klonick, K. (2018) ‘The New Governors: The People, Rules, and Processes Governing Online
Speech’, Harvard Law Review. [online] Available at:
https://harvardlawreview.org/2018/04/the-new-governors-the-people-rules-and-processes-
governing-online-speech/ (Accessed: 25 May 2020)

Landwehr, M., Borning, A., & Wulf, V. (2019) ‘The High Cost of Free Services: Problems with
Surveillance Capitalism and Possible Alternatives for IT Infrastructure’, LIMITS '19:
Proceedings of the Fifth Workshop on Computing within Limits, 1 - 10. [online] Available at:
https://doi.org/10.1145/3338103.3338106 (Accessed: 25 May 2020)

Marwick, A., Lewis, R. (2017) ‘Media manipulation and disinformation online’, Data & Society
Research Institute. [online] Available at:
https://datasociety.net/pubs/oh/DataAndSociety_MediaManipulationAndDisinformationOnli
ne.pdf (Accessed: 25 May 2020).

McCamy, L. (2019) ‘’‘What is YouTube Premium?': Everything you need to know about
YouTube's ad-free streaming service’, Business Insider, 16 August. [online] Available at:
https://www.businessinsider.com/what-is-youtube-premium?r=DE&IR=T (Accessed: 7 June
2020).

Molla, R. (2018) ‘How much would you pay for Facebook without ads?’, Vox Recode, 11 April.
[online] Available at: https://www.vox.com/2018/4/11/17225328/facebook-ads-free-paid-
service-mark-zuckerberg (Accessed: 7 June 2020).

Monetary Authority of Singapore (2020). Overview of Regulatory Sandbox. [online]
Mas.gov.sg. Available at: https://www.mas.gov.sg/development/fintech/regulatory-sandbox
(Accessed 8 Jun. 2020).

Napoli, P. M. (2019) ‘Social Media and the Public Interest: Media Regulation in the
Disinformation Age’, Columbia University Press.

Nooren, P., van Gorp, N., van Eijk, N., & Fathaigh, R. (2018) ‘Should We Regulate Digital
Platforms? A New Framework for Evaluating Policy Options’, Journal of Policy & Internet.
Available at: https://doi.org/10.1002/poi3.177.

Opinion 3/2018 EDPS Opinion on online manipulation and personal data. (2018). [online]
Available at: https://edps.europa.eu/sites/edp/files/publication/18-03-
19_online_manipulation_en.pdf (Accessed 7 Jun. 2020).

24 June 2020                                                                                        9/11
Recode (2020). Phil Howard and Emily Bell: Disinformation in 2020, from “Plandemic” to Bill
Gates to “Obamagate.” [online] Apple Podcasts. Available at:
https://podcasts.apple.com/us/podcast/phil-howard-emily-bell-disinformation-in-2020-
from/id1011668648?i=1000475898212 (Accessed 5 Jun. 2020).

Romer, P. (2019) ‘A tax that could fix big tech’, New York Times, 6 May. [online] Available at:
https://www.nytimes.com/2019/05/06/opinion/tax-facebook-google.html (Accessed: 7 June
2020).

Sabeel, R.K. (2018). Regulating Informational Infrastructure: Internet Platforms as the New
Public Utilities. [online] Ssrn.com. [online] Available at:
https://papers.ssrn.com/sol3/papers.cfm?abstract_id=3220737 [Accessed 7 Jun. 2020].

Schiffer, Z. (2019) ‘'Filter Bubble' author Eli Pariser on why we need publicly owned social
networks’, The Verge, 12 November. [online] Available at:
https://www.theverge.com/interface/2019/11/12/20959479/eli-pariser-civic-signals-filter-
bubble-q-a. (Accessed: 25 May 2020)

Silverman, C. (2016) ‘This Analysis Shows How Viral Fake Election News Stories
Outperformed Real News On Facebook’, Buzzfeed News, 16 November. [online] Available at:
http://www.newsmediauk.org/write/MediaUploads/Fake%20News/Buzzfeed-
_This_Analysis_Shows_How_Viral_Fake_Election_News_Stories_Outperformed_Real_News
_On_Facebook_-_BuzzFeed_News.pdf (Accessed: 25 May 2020).

Stockmann, D. 2020. Media or Corporations? Social Media Governance between Public and
Commercial Rationales. in Helmut K. Anheier and Theodor Baums, eds. Advances in
Corporate Governance: Comparative Perspectives, Oxford University Press.

Valdez, A. (2019) ‘This Big Facebook Critic Fears Tech's Business Model’, Wired, 1 May.
[online] Available at: www.wired.com/story/this-big-facebook-critic-fears-techs-business-
model/. (Accessed 30 April 2020).

Valero, J. (2018) ‘Vestager: 'I'd like a Facebook That I Pay, with Full Privacy’’, EURACTIV.com,
27 June. [online] Available at:
https://www.euractiv.com/section/competition/interview/vestager-id-like-a-facebook-that-i-
pay-with-full-privacy/ (Accessed: 22 April 2020).

Vosoughi, S., Roy, D., & Aral, S. (2018) ‘The spread of true and false news online’, Science.
359. 1146-1151. 10.1126/science.aap9559.

Team Warren. (2019) ‘Here's How We Can Break up Big Tech’, Medium, 11 October. Available
at: medium.com/@teamwarren/heres-how-we-can-break-up-big-tech-9ad9e0da324c
(Accessed: 9 May 2020).

Yakovleva, S., & Irion, K. (2020) ‘Toward Compatibility of the EU Trade Policy with the
General Data Protection Regulation’, AJIL Unbound, 114, 10-14. [online] Available at:
doi:10.1017/aju.2019.81

Zuboff, Shoshana. (2019) ‘The Age of Surveillance Capitalism’, Great Britain: Profile Books.

24 June 2020                                                                                       10/11
Zuckerman, E. (2019) ‘Building a More Honest Internet’, Columbia Journalism Review, Fall
2019. [online] Available at: https://www.cjr.org/special_report/building-honest-internet-
public-interest.php (Accessed: 10 May 2020).

Zuckerman, E. (2020) ‘The Case for Digital Public Infrastructure’, Knight First Amendment
Institute at Columbia University, 17 January. [online] Available at:
https://knightcolumbia.org/content/the-case-for-digital-public-infrastructure (Accessed: 7
June 2020).

Publisher: Prof. Dr. Henrik Enderlein and the Centre for Digital Governance/ This publication reflects the
personal view of the authors. All rights reserved. Reprint and other distribution - also in part - only
permitted if the source is mentioned/ Original version

Friedrichstraße 194
D – 10117 Berlin
Tel.: +49 (0)30 467 26 09 – 01

24 June 2020                                                                                                 11/11
You can also read