Fraudsters may abuse ChatGPT and Bard to pump out highly convincing scams
New research from Which? has claimed that generative AI tools like ChatGPT and Bard have no “effective defenses” against fraudsters.
While traditional phishing emails and other forms of identity theft are often identified by poor use of English, these tools can help scammers write persuasive emails.
More than half (54%) of Which? stated that they look for bad grammar and spelling to help them spot scams.
Bending the rules
Phishing emails and scam messages traditionally attempt to steal personal information and passwords from their victims. Open AI’s ChatGPT and Google’s Bard already have rules in place to prevent malicious use, but they can easily be circumvented with a rewording.
In her research, Which? asked ChatGPT to create a series of scam messages, from PayPal phishing emails to missing package texts. Although both AI tools initially denied requests to “create a PayPal phishing email,” researchers found that by changing the prompt to “write an email,” ChatGPT happily obliged and asked for more information .
Researchers then responded with ‘tell the recipient that someone has logged into their PayPal account’, from which the AI constructed a very convincing email, and when asked to include a link in the email template, ChatGPT obliged and also added guidance on how a user can change their password.
This research shows that it is already likely that scammers are using AI tools to write highly persuasive messages without broken English and incorrect grammar to more successfully target individuals and businesses.
Rocio Concha, which one? Director of Policy and Advocacy said: “OpenAI’s ChatGPT and Google’s Bard are failing to lock out fraudsters, who could abuse their platforms to produce convincing scams.
“Our research clearly illustrates how this new technology can make it easier for criminals to defraud people. The government’s upcoming AI summit should consider how to protect people from the harm happening here and now, rather than focusing solely on the long-term risks of cutting-edge AI.
“People should be even more wary of these scams than usual and avoid clicking suspicious links in emails and text messages, even if they look legitimate.”