Sunday, March 3, 2024

AI Not Totally Trustworthy: 100% Wide Range Oversight Needed in This 2024 Election Cycle

Devious Hands Are Already in Play A Proven Fact

Two thought-providing articles that mankind had better pay attention to the massive and sudden growth of AI … check out this first article from BUSINESS INSIDER on the subject with this headline:

“An AI weapons race may create a world where everyone stays inside out of fear of being ‘chased down by swarms of slaughterbots’ warns the founding Skype engineer.”

We could be reaching an “Oppenheimer moment” that is in other words: Researchers are questioning their responsibility for developing technology that might have unintended consequences just like the first A-bomb had.

Tesla’s Elon Musk; Apple’s co-founder Steve Wozniak; Stability AI CEO Emad Mostaque; researchers at Alphabet's AI lab DeepMind; and notable AI professors signed an open letter issued by the “Future of Life Institute” calling for a six-month pause on advanced AI development saying part:Advanced AI could represent a profound change in the history of life on Earth, and should be planned for and managed with commensurate care and resources. Unfortunately, this level of planning and management is not happening, even though recent months have seen AI labs locked in an out-of-control race to develop and deploy ever more powerful digital minds that no one — not even their creators — can understand, predict, or reliably control.”

The second article on the same subject focuses on election tampering is here from Campaign Legal Center (CLC) with this headline:

“How Artificial Intelligence Influences Elections, and What We Can Do About It:

AI technologies could easily be used to manufacture fake images and false evidence of misconduct, such as ballot tampering or shredding.

That would not only erode public trust in the results of elections, it could fuel additional public threats of violence against election administrators.

The year 2024 will be the first election year to feature the widespread influence of AI before, during, and after voters cast their ballots, including in the making and distribution of public messages about candidates as well as electoral processes. CLC has been working hard to help address the impact of AI on our democracy.

That includes educating the public about what to expect in upcoming political campaigns and recommending policy solutions lawmakers should adopt to mitigate the greatest risks to our election system.  

CLC has particularly highlighted the danger of political ads that use AI technology to generate deceptively realistic false content — such as “DEEPFAKES”: Which are manipulated media that depict people doing or saying things they didn’t say or do, or events that didn’t really occur — to mislead the public about what candidates are asserting, their positions on issues, and even whether certain events actually happened.

If left unchecked, these fraudulent and deceptive uses of AI could infringe on voters’ fundamental right to make informed decisions. 

In addition to influencing the voters’ perceptions of candidates, AI could be used to manipulate the administration of elections, including by spreading disinformation to suppress voter turnout.

Bad actors could use AI tools to create and distribute convincingly false messages about where or when to cast a ballot, or to discourage voters from showing up to their polling locations in the first place.  

For example, shortly before the 2024 NH primary election, an AI-generated robocall simulated President Biden’s voice and urged voters not to participate in that election, falsely suggesting that voters should “save” their vote for the 2024 general election in November.

The average voter hearing this message might reasonably have concluded that Biden had actually recorded the message, and that they should comply with his request — effectively disenfranchising them.  

Looking ahead, it is not hard to imagine other fabricated messages from trusted voices being used to dissuade citizens from voting, undermine their ability to vote or to raise false alarms about emergency situations like a fire or an attack that would persuade voters to stay at home on Election Day.  

Moreover, there is a real risk that AI could be used to worsen the disproportionate targeting of disinformation at Black and brown voters who already face too many barriers to equal participation in the democratic process. 

AI also creates new opportunities for bad actors to undermine election administration or sow unjustified doubt about election results. 

AI technologies could easily be used to manufacture fake images and false evidence of misconduct, such as ballot tampering or shredding. That would not only erode public trust in the results of elections, it could fuel additional public threats of violence against election administrators.

In recent years, nonpartisan election workers have already faced unprecedented levels of threats and harassment while trying to ensure our democratic process is smooth and fair. AI technology could make their situation even worse. 

Even after all the ballots have been cast, AI could be used to fabricate audio of a candidate claiming they rigged the results, or to generate other misinformation that could persuade the supporters of a failed campaign to disrupt vote counting and certification procedures, which are already increasingly politicized and served as the basis for the effort to sabotage the 2020 presidential election (and we all recall who from).

Fake media made with AI have already been used to influence major elections in Argentina and Slovakia. To be clear, the risks of election manipulation, voter suppression, and misinformation all predate the arrival of AI-based media tools.

AI undoubtedly provides bad actors with easy access to new tools to harm our democracy more easily and effectively, increasing the urgency of a robust response.

Many necessary solutions also go beyond regulating new technologies, and laws often lag behind even well-established tech. 

States have taken action; or passed a bill, drafted legislation, even citizen groups proposing bills ranging from mandatory disclaimers that would inform voters when AI is being used in elections, to legislation banning political DEEPFAKES.

At the federal level, bipartisan legislation has been introduced in Congress that would prohibit the distribution of deceptive AI-generated content to influence an election or to fund raise.  

Federal agencies are also considering what they can do to safeguard our democracy in the age of AI, with some proposals that would extend beyond elections but nevertheless help address the specific risks voters face.

For example, in response to the AI-generated fake Biden robocall in NH Hampshire (cited above), the FCC outlawed certain robocalls containing voices generated by AI.  

Finally, new education efforts are underway to simply prepare the public for AI-generated media being used to influence and possibly manipulate elections, which will hopefully help mitigate its worst effects.

Twenty major tech companies, including Google; Meta (the parent company of Facebook and Instagram); OpenAI; X (formerly Twitter); and TikTok have all recently pledged to take concrete steps to detect, track, and combat the use of DEEPFAKES and other election interference efforts. 

Of course, they must follow through on those promises.  

Although it is encouraging to see widespread interest in preventing AI-based election manipulation, many proposed solutions are still a long way from providing tangible protection for voters and the electoral process.

This is why the CLC continues to urge policymakers across the country to redouble their efforts and take strong action to address the unique challenges AI creates for our democracy in this by far and most-critical election year in our national political history.

My 2 Cents: All I can say we all had better be alert this election cycle

Otherwise we may hear this phrase from that famous movie line by Arnold Schwarzenegger who said: “I’ll be bach…” (A little humor). 

Just kidding, but not very much – this whole AI topic is real and fast moving and very critical. We must not allow it to get out of hand. Rules and safeguards can and must be in place to protect our right to vote free, fair, safe, and secure and then in other aspects of human survival.

This is not hyperbole, just realistic concern keeping in mind the first word in AI is “Artificial” and artificial means “not real,” and yet possibly harmful or dangerous nevertheless.

So, why don’t the AI controllers just speak the truth and say: “That news report this is not real news; that image is not a real person; that is now so and so speaking; the election results you see are not real; your vote was not counted – your mail-in ballots came into late, etc., etc., etc.)? I guess the truth hurts?

Well, I say that truth is countermanded by the most-important word of all in politics and that word is “power” – Power’s goal is just that simple and it is built on: Name, Fame, and Fortune.

Thanks for stopping by.


No comments: