Audio By Carbonatix
When we talk about the dangers posed by artificial intelligence, the emphasis is usually on the unintended side effects.
We worry that we might accidentally create a super-intelligent AI and forget to program it with a conscience; or that we’ll deploy criminal sentencing algorithms that have soaked up the racist biases of their training data.
But this is just half the story.
What about the people who actively want to use AI for immoral, criminal, or malicious purposes? Aren’t they more likely to cause trouble — and sooner? The answer is yes, according to more than two dozen experts from institutes including the Future of Humanity Institute, the Centre for the Study of Existential Risk, and the Elon Musk-backed non-profit OpenAI. Very much yes.
In a report published today titled “The Malicious Use of Artificial Intelligence: Forecasting, Prevention, and Mitigation,” these academics and researchers lay out some of the ways AI might be used to sting us in the next five years, and what we can do to stop it. Because while AI can enable some pretty nasty new attacks, the paper’s co-author, Miles Brundage of the Future of Humanity Institute, tells The Verge, we certainly shouldn’t panic or abandon hope.
“I like to take the optimistic framing, which is that we could do more,” says Brundage. “The point here is not to paint a doom-and-gloom picture — there are many defenses that can be developed and there’s much for us to learn. I don’t think it’s hopeless at all, but I do see this paper as a call to action.”
The report is expansive but focuses on a few key ways AI is going to exacerbate threats for both digital and physical security systems, as well as create completely new dangers. It also makes five recommendations on how to combat these problems — including getting AI engineers to be more upfront about the possible malicious uses of their research; and starting new dialogues between policymakers and academics so that governments and law enforcement aren’t caught unawares.
Let’s start with potential threats, though: one of the most important of these is that AI will dramatically lower the cost of certain attacks by allowing bad actors to automate tasks that previously required human labor.
Take, for example, spear phishing, in which individuals are sent messages specially designed to trick them into giving up their security credentials. (Think: a fake email from your bank or from what appears to be an old acquaintance.) AI could automate much of the work here, mapping out an individuals’ social and professional network, and then generating the messages. There’s a lot of effort going into creating realistic and engaging chatbots right now, and that same work could be used to create a chatbot that poses as your best friend who suddenly, for some reason, really wants to know your email password.
This sort of attack sounds complex, but the point is that once you’ve built the software to do it all, you can use it again and again at no extra cost. Phishing emails are already harmful enough — they were responsible for both the iCloud leak of celebrity’s pictures in 2014, as well as the hack of private emails from Hillary Clinton’s campaign chairman John Podesta. The latter not only had an influence on the 2016 US presidential election, it also fed a range of conspiracy theories like Pizzagate, which nearly got people killed. Think about what an automated AI spear-phisher could do to tech-illiterate government officials.
The second big point raised in the report is that AI will add new dimensions to existing threats. With the same spear phishing example, AI could be used to not only generate emails and text messages, but also fake audio and video. We’ve already seen how AI can be used to mimic a target’s voice after studying just a few minutes of recorded speech, and how it can turn footage of people speaking into puppets. The report is focused on threats coming up in the next five years, and these are fast becoming issues.
And, of course, there is a whole range of other unsavory practices that AI could exacerbate. Political manipulation and propaganda for a start (again, areas where fake video and audio could be a huge problem), but also surveillance, especially when used to target minorities. The prime example of this has been in China, where facial recognition and people-tracking cameras have turned one border region, home to the largely Muslim Uighur minority, into a “total surveillance state.”
These are just examples of AI’s capacity to scale becoming a threat. It replaces the humans who watch the feeds, turning CCTV camera from passive into active observers, allowing them to categorize human behavior automatically. “Scalability in particular is something that hasn’t got enough attention,” says Brundage. “It’s not just the fact that AI can perform at human levels at certain tasks, but that you can scale it up to a huge number of copies.”
Finally, the report highlights the entirely novel dangers that AI creates. The authors outline a number of possible scenarios, including one where terrorists implant a bomb in a cleaning robot and smuggle it into a government ministry. The robot uses its built-in machine vision to track down a particular politician, and when it’s near, the bomb detonates. This takes advantage of new products AI will enable (the cleaning robots) but also its autonomous functions (the machine vision-based tracking).
Outlining scenarios like this may seem a bit fantastical, but we’ve really already begun to see the first novel attacks enabled by AI. Face-swapping technology has been used to create so-called “deepfakes” — pasting the faces of celebrities onto pornographic clips without their consent. And although there have been no high-profile cases of this to date, we know those involved in creating this content want to test it out on people they know; creating perfect fodder for harassment and blackmail.
These examples only take in a portion of the report, but the whole document leaves you wondering: what’s to be done? The solutions are easy to outline, but will be challenging to follow through on. The report makes five key recommendations:
AI researchers should acknowledge how their work can be used maliciously
Policymakers need to learn from technical experts about these threats
The AI world needs to learn from cybersecurity experts how to best protect its systems
Ethical frameworks for AI need to be developed and followed
And more people need to be involved in these discussions. Not just AI scientists and policymakers, but also ethicists, businesses, and the general public
In other words: a little more conversation and a little more action please.
It’s a big ask considering what a complex and nuanced subject artificial intelligence is, but there have been promising signs. For example, with the rise of deepfakes, web platforms reacted quickly, banning the content and stopping its immediate spread. And lawmakers in the US have already started talking about the problem — showing that these debates will reach government if they’re urgent enough.
“There’s certainly interest,” says Brundage of government involvement in discussing these topics. “But there’s still a sense that more discussion needs to happen in order to find out what are the most critical threats, and what are the most practical solutions.” And in most cases, he says, it’s difficult to even judge what will be a threat when. “It’s unclear how gradual all this will be — whether there’ll be a big catastrophic event, or whether it’ll be a slow rolling thing that gives us plenty of opportunities to adapt.”
“But that’s exactly why we’re raising these issues now.”
DISCLAIMER: The Views, Comments, Opinions, Contributions and Statements made by Readers and Contributors on this platform do not necessarily represent the views or policy of Multimedia Group Limited.
Tags:
DISCLAIMER: The Views, Comments, Opinions, Contributions and Statements made by Readers and Contributors on this platform do not necessarily represent the views or policy of Multimedia Group Limited.
Latest Stories
-
Kpandai re-run: Supreme Court decision only defers NPP’s defeat – Tanko-Computer
15 minutes -
Annoh-Dompreh demands Ministerial reshuffle as absences stall Parliament
19 minutes -
Police arrest 2 suspects after viral child abuse video in Agona Jamasi
36 minutes -
Kloma Gbi takes ‘Taking Them Along’ career mentorship programme to Asesewa
38 minutes -
2026 World Cup: Nketiah, Hudson-Odoi will make Black Stars stronger – Former Sports Minister
43 minutes -
VAST Ghana demands immediate ban on alcohol industry sponsorship of school events
47 minutes -
Be proactive with changing trends in regulatory and tax landscape – Deloitte Partner to businesses
48 minutes -
Supreme Court order only suspends Kpandai rerun, does not reinstate Nyindam – Defeamekpor
51 minutes -
Aquaculture players validate framework for Blue Food Innovation Hub
58 minutes -
JoyBusiness Van: Rose Aluminium moulds Ghana’s cookware industry
1 hour -
Atiwa East DCE must resign or be sacked – Senyo Hosi
1 hour -
Petition for removal of Justice Yoni Kulendi dismissed
1 hour -
Stanbic Asantehene Invitational Golf Championship draws elite field of golfers
2 hours -
New Year debt recovery school launched to enhance professional competence in Ghana’s financial landscape
2 hours -
Lightning kills mother and 4-year-old son at Akyeremade
2 hours
