OUR PARTNERS

States Take Action On Artificial Intelligence In Political Advertising


19 June, 2024

The increasing prevalence of artificial intelligence in our daily lives presents both opportunities and challenges, particularly in the realm of political discourse. One emerging challenge is the regulation of “deep fakes” within political advertising, a topic that has attracted the attention of lawmakers across the United States.

Deep fakes, which leverage AI to create highly convincing and often indistinguishable video and audio manipulation, are a growing concern within the political arena. The potential for misuse to spread misinformation and manipulate public perception is immense. To address these concerns, approximately 17 states have enacted legislation, with most requiring labeling of deep fakes in ads, while a few have taken the additional step of banning them outright.

Federally, the tide is beginning to turn as well. The Federal Communications Commission (FCC) and the Senate Committee on Rules and Administration have both stepped up with proposals aimed at safeguarding the integrity of political communications.

The FCC’s recent proposal, as outlined by Chairwoman Jessica Rosenworcel, suggests mandatory disclaimers when AI images generator technology or other AI tools are used in political ads. The move would extend to broadcasters, local cable, and other media obliged to follow FCC regulations, ensuring that the use of AI is transparent to viewers and that appropriate records are maintained in publicly accessible FCC files.

However, the notion of regulation brings its own set of complexities, particularly around defining what constitutes AI-generated content. AI’s broad spectrum encompasses simple video editing tools as well as advanced AI video generator programs capable of creating entirely fictitious scenarios. The challenge for legislators is to draft language that clearly delineates the use cases that necessitate regulation without stifling legitimate and innocuous uses of AI, such as video production enhancements that do not materially misinform the public.

These regulatory initiatives must also consider exceptions such as satire, parody, or journalistic uses that serve the public interest. Ensuring these nuances are captured demands careful crafting of legislation at both the state and federal levels.

Perhaps the trickiest aspect of regulating deep fakes is the enforcement mechanism. Current technology for recognizing AI-created content can be unreliable, raising concerns about broadcasters’ ability to accurately identify deep fakes in a timely and accurate manner. The risk of over-regulating and inadvertently suppressing legitimate political speech is a real concern, with the potential for improperly flagged ads causing unnecessary censorship.

The enforcement burden on broadcasters is substantial, and the consequences of failing to identify deep fakes pose legal risks. The tools available for detection have been scrutinized for their potential inaccuracies in real-world scenarios, as highlighted in the latest ai news articles. For instance, high-end detection systems often fall short in the face of evolving deep fake technology, while the resource-intensive process of accurate verification may take weeks—an impractical timeframe given the rapid pace of political campaigns.

Moreover, such responsibilities imposed on local broadcasters and cable operators could be prohibitive, given that many lack the means to employ state-of-the-art AI text generator identification tools that large corporations might access.

The debate over AI-generated political content is not about stifling the innovative progress of technology; rather, it’s about preserving the cornerstone of democracy: informed decision-making by the electorate. Striking a balance between the use of cutting-edge ai tools and maintaining public trust in political messaging remains an imperative task for policymakers.

As we await the text of the FCC’s proposed rules and track the progress of relevant legislation, stakeholders from tech, media, and regulatory bodies should engage in candid discussions about realistic and fair approaches to AI regulation in political advertising. Transparency, accuracy, and accountability in AI-generated political content are essential to ensuring a fair electoral process and maintaining the integrity of democracy. With thoughtful regulation and responsible AI usage, we can hope to uphold these principles in the digital age.