View this post on the web at [link removed]
Last week, the Federal Communications Commission banned robocalls [ [link removed] ] that use voices generated by artificial intelligence (AI). The action followed a series of calls placed in New Hampshire that attempted to dissuade people from voting in the state’s primary—using the replicated voice of President Joe Biden.
The incident in New Hampshire is one of the most egregious uses of AI generated deepfakes indirectly influencing an election. Unfortunately, this isn’t new. Deepfake technology has made headlines since at least 2019, when the tools were used to target then-Speaker of the House Nancy Pelosi [ [link removed] ], Rep. Alexandria Ocasio-Cortez [ [link removed] ] of New York, and former President Donald Trump [ [link removed] ]. AI has become so pervasive, it’s even found its way into comedy sketches [ [link removed] ].
But it’s not all in good fun. Experts are concerned about how AI will impact elections [ [link removed] ], creating a situation where false information is believed to be true, and true information is believed to be false. The possibilities scream for government regulation, but the challenge is to do it in a way that doesn’t stunt a sector with tremendous potential for human advancement. With the presidential election looming, the clock is ticking. —Melissa Amour, Managing Editor
AI’s Threat to Democracy
Just before a recent election in Slovakia, a fake audio recording [ [link removed] ] in which a candidate boasted about rigging the election went viral. The incident raised alarms in the US, where it’s not difficult to envision a similar scenario playing out in an election year.
AI is already being used to craft highly targeted and believable phishing campaigns [ [link removed] ] and can also be deployed in cyberattacks [ [link removed] ] that compromise election security. None of these threats are new; we’ve seen similar activity in 2016 [ [link removed] ] and 2020 [ [link removed] ]. But advancements in AI [ [link removed] ] since those elections and the widespread availability of the technology can significantly aid bad actors in 2024.
The threats to democracy posed by AI are ubiquitous, and the government [ [link removed] ] must implement safeguards against uses of AI that could intentionally deceive the public [ [link removed] ]. The challenge is whether it can restrain the technology effectively without stifling it completely. It’s a tall order. Let’s take a look at where things stand.
What You Should Know
Proposed AI legislation thus far largely addresses generative AI [ [link removed] ], and protecting the public from manipulated content as it relates to personal privacy. To that end, a bipartisan group in Congress introduced the No Artificial Intelligence Fake Replicas And Unauthorized Duplications (No AI FRAUD) Act last month.
“[T]he No AI FRAUD Act, will punish bad actors using generative AI to hurt others—celebrity or not. Everyone should be entitled to their own image and voice and my bill seeks to protect that right." —Rep. María Elvira Salazar [ [link removed] ] of Florida
But hopes are not high [ [link removed] ] that Congress will pass any AI-related legislation before the election, so states and cities are taking individual regulatory action [ [link removed] ]. As of September, state legislatures had introduced 191 AI-related bills [ [link removed] ] in 2023—a 440% increase over 2022.
“States being the laboratories of democracy, they can act a little bit more quickly…and some are, both through their boards or organizations of elections or through state legislation. But this is a space that we certainly have to come to some formal agreement on because again, it is about protecting democracy.” —US Rep. Marcus Molinaro [ [link removed] ] of New York
The White House’s executive action on AI has been the most comprehensive [ [link removed] ] of all, and the administration has released a blueprint [ [link removed] ] for an AI Bill of Rights as well.
“The Order directed sweeping action to strengthen AI safety and security, protect Americans’ privacy, advance equity and civil rights, stand up for consumers and workers, promote innovation and competition, advance American leadership around the world, and more.” —White House AI Fact Sheet [ [link removed] ]
But it’s a fine line. Lawmakers must craft regulation that doesn’t strangle innovation in their zeal to protect democracy, jobs, and humanity itself. In short, we don’t want to lose our competitive edge through over-regulation, like Europe [ [link removed] ], and allow competitors like China [ [link removed] ] to take the lead.
How We Got Here
Our hopes and fears for the future of AI are fundamentally shaped by our recent experience with the rise of social media. It is an obvious analogy, similarly new, destabilizing, enduring, and to some extent inevitable. Many lawmakers [ [link removed] ] on both sides of the aisle believe insufficient regulation of social media [ [link removed] ] led to unintended consequences detrimental to democracy and the general well being of individuals.
“Because different sides see different facts [on social media], they share no empirical basis for reaching a compromise. Because each side hears time and again that the other lot are good for nothing but lying, bad faith, and slander, the system has even less room for empathy. Because people are sucked into a maelstrom of pettiness, scandal, and outrage, they lose sight of what matters for the society they share.” —The Economist [ [link removed] ]
But on the other hand, the US is the undisputed leader in social media. The largest social media companies are headquartered in the US, creating jobs and generating economic growth. As a result, the US has marginally more control of the sector globally as the chief regulatory body.
“[T]he internet economy’s contribution to the US GDP grew 22 percent per year since 2016, in a national economy that grows between two to three percent per year.”—Interactive Advertising Bureau/Harvard Business School study [ [link removed] ]
Tech innovation is at the forefront of modern American economic success.
“The seven largest US technology companies, often referred to as the ‘Magnificent Seven’ and comprising Alphabet Inc (Google), Amazon.com Inc, Apple Inc, Meta Platforms Inc, Microsoft Corp, Nvidia Corp, and Tesla Inc, boast a combined market value exceeding a staggering $12 trillion (€11 trillion). To put this in perspective, their value is almost equivalent to the combined gross domestic products of the four largest European economies: Germany, the United Kingdom, France, and Italy, which amount to $13 trillion.” —EuroNews [ [link removed] ]
Government leaders [ [link removed] ] are determined to generate safer outcomes with AI through greater engagement with tech companies [ [link removed] ], rather than relying on them to police themselves which has produced, at best, a spotty record. But as our experience with social media shows, regulation can—and should—only go so far, lest it diminish economic opportunity and US influence internationally.
What People Are Saying
One thing is certain: AI will change democracy [ [link removed] ], in the same way that previous technologies, like social media, the internet, and television, did before it. The challenge is how to regulate it to ensure those changes are largely advantageous [ [link removed] ], without stifling the development of a groundbreaking emergent technology. Some people are worried that proposed regulations could go too far.
“[P]re-emptive regulation can erect barriers to entry for companies interested in breaking into an industry. Established players, with millions of dollars to spend on lawyers and experts, can find ways of abiding by a complex set of new regulations, but smaller start-ups typically don’t have the same resources. This fosters monopolization and discourages innovation.” —Tim Wu [ [link removed] ], law professor at Columbia University and author of “The Curse of Bigness: Antitrust in the New Gilded Age.”
Meanwhile others’ fears about AI’s disruptive or even apocalyptic [ [link removed] ] potential are driving calls for stringent measures.
“Bad actors using AI…represents a dire threat to the very core of democratic societies. As AI technology continues to advance, the potential for abuse grows, making it imperative to act decisively. Strengthening legal frameworks, promoting ethical AI, increasing public awareness, and fostering international cooperation are critical steps in combating this threat.” —Neil Sahota [ [link removed] ], IBM Master Inventor, United Nations AI advisor, and author of “Own the AI Revolution”
In the end, a balanced approach that minimizes harm but encourages innovation seems ideal. After all, AI could lead to overwhelming advances in human knowledge and ingenuity.
“The path ahead demands robust legal frameworks, respect for intellectual property, and stringent safety standards, akin to the meticulous oversight of nuclear energy. But beyond regulations, it requires a shared vision. A vision where technology serves humanity and innovation is balanced with ethical responsibility. We must embrace this opportunity with wisdom, courage, and a collective commitment to a future that uplifts all of humanity.” —Former US Rep. Will Hurd [ [link removed] ] of Texas, who previously served as a board member of OpenAI
AI, democracy, and disinformation: How to stay informed without being influenced by fake news —The DePaulia [ [link removed] ]
Inside OpenAI's plan to make AI more 'democratic' —Time [ [link removed] ]
‘One-size-fits-all’ approach not fit for deepfakes —Economic Times [ [link removed] ]
Opinion: Effective AI regulation requires understanding general-purpose AI —Brookings [ [link removed] ]
Opinion: Bridging the trust gap: Using AI to restore faith in democracy —Federal Times [ [link removed] ]
‘Un-American’: Donald Trump threatens NATO, says he might ‘encourage’ Russian attack —USA Today [ [link removed] ]
Flipped seat in New York: Democrat Tom Suozzi wins special election to replace ousted US Rep. George Santos —NBC News [ [link removed] ]
Second time’s the charm: House Republicans impeach Homeland Security Secretary Alejandro Mayorkas in historic, controversial vote —ABC News [ [link removed] ]
Food for thought: Report shows 23.5M independent voters are disenfranchised in 2024 primaries —Unite America [ [link removed] ]
Give this a listen: Ukraine aid on the line, Tucker goes to Moscow —The Dispatch [ [link removed] ]
Holland: The challenge to Holland’s liberalism —The UnPopulist [ [link removed] ]
Hungary: Prime Minister Viktor Orbán faces political crisis after president’s resignation —PBS [ [link removed] ]
Lebanon: Israel launches strikes in Lebanon after rockets hit army base —BBC [ [link removed] ]
Russia: Warning from House Intel about Russia’s space power —Politico [ [link removed] ]
Ukraine: NATO's Stoltenberg urges US House to pass Ukraine aid, says China watching —Reuters [ [link removed] ]
Hey Topline readers, you remember the drill. We want to hear your reactions to today’s stories. We’ll include some of your replies in this space in our next issue of The Topline. Click here [ [link removed] most current link. ] to share your take, and don’t forget to include your name and state. We’re looking forward to hearing from you!
Thanks for reading The Topline! Subscribe for free to receive new posts and support my work.
Unsubscribe [link removed]?