Twenty technology companies have made a commitment to prevent their artificial intelligence (AI) software from influencing elections, including those in the United States. The companies recognize the significant risk their products pose, especially in a year when around 4 billion people are expected to participate in elections worldwide. The agreement raises concerns about deceptive AI election content and its potential to mislead the public, thereby posing a threat to the integrity of electoral processes. It also acknowledges the slow response of global lawmakers to the rapid progress in AI and highlights the tech industry’s exploration of self-regulation.
Brad Smith, Vice Chair and President of Microsoft, supports the idea of preventing the weaponization of AI in elections by saying, “As society embraces the benefits of AI, we have a responsibility to help ensure these tools don’t become weaponized in elections.” The 20 signatories of the pledge include major tech companies like Microsoft, Google, Adobe, Amazon, and IBM, among others. It is important to note that the agreement is voluntary and does not impose a complete ban on AI content in elections.
The document outlines eight steps that the companies commit to taking this year. These steps include developing tools to distinguish AI-generated images from genuine content and promoting transparency by sharing significant developments with the public. Despite the commitments made, open internet advocacy group Free Press argues that tech companies have not followed through with previous pledges for election integrity after the 2020 election. They advocate for increased oversight by human reviewers.
Congresswoman Yvette Clarke of the 9th District of New York welcomes the tech accord and hopes that Congress will build upon it. She emphasizes the importance of the agreement, stating, “This could be a defining moment for this Congress, and this may be the one unifying issue where we can band together to protect this nation and future generations of Americans to come.” Clarke has introduced legislation to regulate deepfakes and AI-generated content in political ads.
On January 31, the Federal Communications Commission (FCC) voted to outlaw AI-generated robocalls that use AI-generated voices. This decision came after a fake robocall claiming to be from President Joe Biden fabricated alarm ahead of January’s New Hampshire primary, highlighting the potential for counterfeit voices, images, and videos in politics. The move by the FCC demonstrates a broader recognition of the need to address the risks associated with AI-generated content and the urgency to take regulatory action.
While this commitment by tech companies is a step in the right direction, it remains to be seen how effectively they will implement the outlined steps. The concerns of Free Press regarding the companies’ track record in fulfilling previous pledges highlight the importance of ongoing oversight and accountability. With the support of lawmakers like Congresswoman Yvette Clarke, there is hope that further regulations will be implemented to protect the integrity of elections from the misuse of AI.
I’m not holding my breath for these tech companies to actually implement the steps outlined in the agreement.
I’m skeptical that they can actually develop effective tools to distinguish AI-generated content.