(BRUSSELS) – EU negotiators agreed Friday on landmark rules to make social media platforms take measures to protect their users from illegal content, goods and services and protect their fundamental rights in the digital sphere.
The provisional political agreement on the Digital Services Act (DSA) sets out an unprecedented new standard for the accountability of online platforms regarding illegal and harmful content.
It provides better protection for internet users and their fundamental rights, and it defines a single set of rules in the internal market, helping smaller platforms to scale up.
The new framework contains EU-wide due diligence obligations that will apply to all digital services that connect consumers to goods, services, or content, including new procedures for faster removal of illegal content as well as comprehensive protection for users’ fundamental rights online.
In scope of the DSA are various online intermediary services. Their obligations under the DSA depend on their role, size, and impact on the online ecosystem. These online intermediary services include:
- Intermediary services offering network infrastructure: Internet access providers, domain name registrars;
- Hosting services such as cloud computing and webhosting services;
- Very large online search engines with more than 10% of the 450 million consumers in the EU, and therefore, more responsibility in curbing illegal content online;
- Online platforms bringing together sellers and consumers such as online marketplaces, app stores, collaborative economy platforms and social media platforms;
- Very large online platforms, with a reach of more than 10% of the 450 million consumers in the EU, which could pose particular risks in the dissemination of illegal content and societal harms.
Concretely, the DSA contains:
- Measures to counter illegal goods, services or content online, such as:
- a mechanism for users to easily flag such content and for platforms to cooperate with so-called ‘trusted flaggers’;
- new obligations on traceability of business users in online market places;
- New measures to empower users and civil society, including:
- the possibility to challenge platforms’ content moderation decisions and seek redress, either via an out-of-court dispute mechanism or judicial redress;
- provision of access to vetted researchers to the key data of the largest platforms and provision of access to NGOs as regards access to public data, to provide more insight into how online risks evolve;
- transparency measures for online platforms on a variety of issues, including on the algorithms used for recommending content or products to users;
- Measures to assess and mitigate risks, such as:
- obligations for very large platforms and very large online search engines to take risk-based action to prevent the misuse of their systems and undergo independent audits of their risk management systems;
- Mechanisms to adapt swiftly and efficiently in reaction to crises affecting public security or public health;
- New safeguards for the protection of minors and limits on the use of sensitive personal data for targeted advertising.
- Enhanced supervision and enforcement by the Commission when it comes to very large online platforms. The supervisory and enforcement framework also confirms important role for the independent Digital Services Coordinators and Board for Digital Services.
The political agreement reached by the European Parliament and the Council is now subject to formal approval by the two co-legislators. Once adopted, the DSA will be directly applicable across the EU and will apply fifteen months or from 1 January 2024, whichever later, after entry into force. As regards the very large online platforms and very large online search engines the DSA will apply from an earlier date, that is four months after their designation.
Fact page on Commission proposal on the Digital Services Act