Subscribe now

Technology

GPT-4 developer tool can be exploited for misuse with no easy fix

OpenAI’s developer tool for its GPT-4 large language model can be misused to trick the AI into providing information to aid would-be terrorists, and fixing the problem won’t be easy

By Jeremy Hsu

1 December 2023

ChatGPT Open AI chat bot on phone screen

AI chatbots can be fine-tuned to provide information that could help terrorists plan attacks

salarko/Alamy

It is surprisingly easy to remove the safety measures intended to prevent AI chatbots from giving harmful responses that could aid would-be terrorists or mass shooters. The discovery seems to be prompting companies, including OpenAI, to develop strategies to solve the problem. But research suggests their efforts have been met with only limited success so far.

OpenAI worked with academic researchers on a so-called…

Sign up to our weekly newsletter

Receive a weekly dose of discovery in your inbox! We'll also keep you up to date with New Scientist events and special offers.

Sign up

To continue reading, subscribe today with our introductory offers

View introductory offers

No commitment, cancel anytime*

Offer ends 28th October 2023.

*Cancel anytime within 14 days of payment to receive a refund on unserved issues.

Inclusive of applicable taxes (VAT)

or

Existing subscribers

Sign in to your account