In News: The European Union (EU) has given final approval to online safety-focused legislation, which is an overhaul of the region’s social media and e-commerce rules, called the Digital Services Act (DSA).
Features of DSA:
- The law tightly regulates the way intermediaries, especially large platforms such as Google, Meta, Twitter, and YouTube, function in terms of moderating user content.
- Faster removals and provisions to challenge: As part of the overhaul, social media companies will have to add “new procedures for faster removal” of content deemed illegal or harmful.
- Social media companies will have to explain to users how their content takedown policy works. The DSA also allows for users to challenge takedown decisions taken by platforms and seek out-of-court settlements.
- VLOPs must allow regulators to access their data to assess compliance and let researchers access their data to identify systemic risks of illegal or harmful content.
- Bigger platforms have greater responsibility: One of the most crucial features of the legislation is that it avoids a one-size fits all approach and places increased accountability on the Big Tech companies. The ‘Very Large Online Platforms’ (VLOPs) and ‘Very Large Online Search Engines’ (VLOSEs), that is platforms, having more than 45 million users in the EU, will have more stringent requirements.
- Direct supervision by European Commission: a key way to ensure that companies do not sidestep the legislation at the member-state level.
- Clearer identifiers for ads and who’s paying for them: Online platforms must ensure that users can easily identify advertisements and understand who presents or pays for the advertisement. They must not display personalised advertising directed towards minors or based on sensitive personal data.
- It will give better protection to users and to fundamental rights online, establish a powerful transparency and accountability framework for online platforms and provide a single, uniform framework across the EU.
- More transparency on how algorithms work: VLOPs and VLOSEs will face transparency measures and scrutiny of how their algorithms work
- Systemic risk analysis by social media companies will drive accountability about the society impacts of their products.
Comparison with Indian Laws:
- In 2021, India had notified extensive changes to its social media regulations in the form of the Information Technology Rules, 2021 (IT Rules)
- IT Rules placed significant due diligence requirements on large social media platforms such as Meta and Twitter.
- These included appointing key personnel to handle law enforcement requests and user grievances, enabling identification of the first originator of the information on its platform under certain conditions, and deploying technology-based measures on a best-effort basis to identify certain types of content.
- Ensuring greater transparency such as if a user has shared child sexual abuse material on a platform.
- However, WhatsApp has filed a case and has alleged that the requirement to identify the first originator of information will dilute the encryption security on its platform and could compromise personal messages of millions of Indians.
- Proposal for further amendments to the IT Rules include the creation of government-backed grievance appellate committees which would have the authority to review and revoke content moderation decisions taken by platforms.
- India is also working on a complete overhaul of its technology policies and is expected to soon come out with a replacement of its IT Act, 2000, which is expected to look at ensuring net neutrality and algorithmic accountability of social media platforms, among other things.
Source: Indian Express