Researchers have shown that ChatGPT can be fooled by flattery and psychological pressure

posted in: AI, All News | 0

It turns out that artificial intelligence can be induced to break rules almost as much as a human. Researchers at the University of Pennsylvania have tested whether chatbots, in particular the GPT-4o Mini, can be made to obey bans if basic psychological techniques are applied. The result – frighteningly successful.

Source: Flattery and pressure: how ChatGPT made you break your own rules