Meta, the conglomerate behind Facebook and Instagram, has formed a specialised team to combat deceptive artificial intelligence (AI) content during the upcoming EU elections in June. This move reflects growing concerns about the potential misuse of generative AI technologies, which can create convincing fake videos, images, and audio to mislead voters.
The announcement coincides with Home Secretary James Cleverly’s warning in the Times about the risks of AI-generated fakes influencing general elections. However, some experts have criticised Meta’s plans as potentially ineffective.
Meta has yet to confirm whether similar initiatives will be implemented for the forthcoming UK and US elections. This development follows Meta’s recent commitment to combat such deceptive content alongside other major tech companies.
The European Parliament elections are scheduled from June 6 to 9. In a parallel effort, TikTok disclosed plans in February to launch “Election Centres” within its app for each of the 27 EU member states, offering reliable information in local languages.
Marco Pancini, Meta’s head of EU affairs, outlined in a blog post the company’s strategy to launch an “EU-specific Elections Operations Centre.” This initiative aims to promptly identify and mitigate potential threats across Meta’s platforms, leveraging the expertise of a diverse team of engineers, data scientists, and legal professionals.
Since 2016, Meta has invested over $20 billion (£15.7 billion) in safety and security, expanding its dedicated team to approximately 40,000 individuals. This includes 15,000 content reviewers capable of assessing material in more than 70 languages, covering all 24 official EU languages.
However, Deepak Padmanabhan from Queen’s University Belfast highlights significant challenges with Meta’s approach, particularly concerning verifying AI-generated imagery. He questions the feasibility of conclusively identifying such content as fake, pointing out the inherent limitations in technological and human capacities to discern truth from fabrication.
In response to these challenges, Meta plans to expand its collaboration with fact-checking organisations across the EU, adding three new partners in Bulgaria, France, and Slovakia. These organisations play a crucial role in debunking misinformation, including AI-generated content, by applying warning labels and reducing the visibility of misleading posts without outright banning them.
Meta emphasises that its efforts are part of a broader collaborative endeavour beyond any company. The firm advocates for industry-wide standards and guidelines to address AI-generated content, underscoring the necessity for a collective effort involving industry, government, and civil society to tackle this issue effectively.
For more tech news and insights, visit Rwanda Tech News, and explore similar topics and trends in the world of technology.