In a move to combat the misuse of AI in elections, OpenAI, the creator of ChatGPT, has released a new strategy. This strategy comes at a crucial time as over 50 nations are gearing up for elections, making this a big year for global democracy.
The rise of AI technologies, particularly ones that are able to create deepfakes, sets a new challenge to voters worldwide.
The Deepfake Dilemma
Deepfake technology, especially tools like OpenAI’s Dall-E, is a major concern.
These tools have the ability to alter images or fabricate new ones, which might allow opponents to depict each other in false and damaging contexts. Even worse, is the fact that these technologies are becoming so advanced that it’s hard to tell what’s real and what isn’t, raising alarms about how they might be able to impact democratic campaigns.
ChatGPT’s Role and Risks
Another aspect of AI technology under scrutiny is text-based generators like ChatGPT. Their ability to produce convincingly human-like writing poses a risk of spreading misinformation. For example, if someone asks ChatGPT about which party is better for students, there’s no guarantee the response will be accurate.
Recognising these threats, OpenAI has articulated a clear approach to address them. As stated in their blog post, the focus is on promoting accurate voting information, implementing well-balanced policies, and enhancing transparency.
More from Artificial Intelligence
- Dropcall Launches UK’s First AI Voicemail App
- AI and the Death of Forgetting: What Are the Implications of AI Remembering Everything?
- How Is Artificial Intelligence Used For Kids’ Toys?
- Experts Share: What Risks Do Investors Face If Superintelligence Never Arrives? N
- Sam Altman Claims We’ve Passed the AI Event Horizon: What Does That Mean?
- What Is A Digital Clone?
- Product Engineering Approaches For Building UX in Generative AI Tools
- Experts Comment: What Do Investors Want To See In Emerging AI Startups?
OpenAI’s Approach
OpenAI has put together a diverse team, including safety systems, threat intelligence, legal, engineering, and policy experts, to investigate and tackle potential abuses of its technology. While OpenAI’s Dall-E tool is designed to avoid generating images of real individuals, other AI startups may lack similar safeguards, making it a broader industry challenge.
Sam Altman, OpenAI’s CEO, has expressed concern over the use of generative AI to disrupt elections, mentioning the possibility of “one-on-one interactive disinformation”. His testimony in Congress highlights how seriously the company is taking these threats.
Proactive Measures and Partnerships
OpenAI has introduced new features in ChatGPT to guide U.S. users to reliable voting information, specifically directing them to CanIVote.org.
Additionally, the team is working on developing new ways to identify AI-generated images. Collaborating with the Coalition for Content Provenance and Authenticity, OpenAI is working on marking fake images with distinctive icons so online users can spot them more easily.
Historical Context and Future Challenges
The use of AI-generated audio in Slovakia’s elections to discredit a candidate shows the real-world implications of these technologies.
As the world prepares for a significant year in democratic exercises, OpenAI’s proactive measures will be crucial in safeguarding the integrity of elections against the potential misuse of AI and deepfake technologies.
The ongoing evolution of these strategies will be key in addressing the challenges posed by AI in the world of politics.
Source: The Independent