The European Commission has launched a formal investigation into Meta, the parent company of Facebook and Instagram, to evaluate its efforts in moderating political content, illegal content, and disinformation on its platforms. This move comes in response to a rise in online pro-Russian propaganda leading up to the EU elections in early June. There are concerns that Meta may not be fulfilling its obligations under the Digital Services Act (DSA), a set of EU regulations designed to create safer online spaces for users.
The investigation is focusing on several key areas where Meta may have fallen short. These include its strategies for combating disinformation campaigns and “coordinated inauthentic behavior” within the EU, as well as the absence of effective third-party tools for monitoring elections and civic discourse in real-time. Of particular worry is the decision to phase out CrowdTangle without providing a suitable alternative.
There is mounting pressure on EU political leaders to address Russian interference in democratic processes across the region, with reports suggesting that nearly every EU country is being targeted by pro-Russian propaganda. The President of the European Commission, Ursula von der Leyen, emphasized the importance of upholding regulations to safeguard democratic processes, especially during election periods. She stressed that major digital platforms like Meta must dedicate sufficient resources to combatting false information.
The European Commission’s inquiry will also examine Meta’s handling of deceptive advertising, its policies regarding the visibility of political content, and the effectiveness of tools for reporting illegal content. EU antitrust chief, Margrethe Vestager, warned about the risks associated with deceptive advertising and the potential impact on online discourse and individual rights. Ensuring trust in online content is crucial to maintaining informed and active citizenship.
As of now, the European Commission has not set a deadline for completing the investigation. If Meta is found to have violated the DSA and fails to address the identified issues, it could face substantial fines amounting to up to 6 percent of its annual turnover. This underscores the seriousness with which regulatory bodies are approaching the issue of content moderation on digital platforms.
The investigation into Meta by the European Commission highlights the growing scrutiny faced by tech giants regarding their responsibilities in moderating online content. As digital platforms play an increasingly central role in shaping public discourse, efforts to combat disinformation and protect democratic processes become paramount. The outcome of this investigation could have far-reaching implications for Meta and set a precedent for other companies in the tech industry.
Leave a Reply