Tech Companies Are Taking Action on AI Election Misinformation. Will it Matter?

  • 📰 TIME
  • ⏱ Reading Time:
  • 40 sec. here
  • 2 min. at publisher
  • 📊 Quality Score:
  • News: 19%
  • Publisher: 53%

France Nouvelles Nouvelles

France Dernières Nouvelles,France Actualités

A mobile phone over a keyboard is seen, with the logo of Meta on its screen.

Wednesday that it would require labels for political ads that have been digitally altered, using AI or other technology, in ways that could be misleading.it was also taking a number of steps to protect elections, including offering tools to watermark AI-generated content and deploying a “Campaign Success Team” to advise political campaigns on AI, cybersecurity, and other related issues.

Often people overestimate the effects of misinformation because they overestimate both how easy it is to change people’s views on charged issues such as voting behavior and how capable misinformation-enabling technologies such as AI are, says Jungherr. This has already happened. In 2019, an allegation that a video of Ali Bongo, then the president of Gabon, was a fake was used toan attempted coup.

Watermarking and provenance measures by AI developers are likely to be ineffective because malicious actors can easily access AI models that have been, such as Meta’s Llama 2, says Jungherr. “I would argue that this is an attempt by these companies to avoid negative coverage,” he says. “I'm not necessarily sure that they expect that these tools will shift an election.”

Nous avons résumé cette actualité afin que vous puissiez la lire rapidement. Si l'actualité vous intéresse, vous pouvez lire le texte intégral ici. Lire la suite:

 /  🏆 93. in FR
 

Merci pour votre commentaire. Votre commentaire sera publié après examen.

France Dernières Nouvelles, France Actualités