Researchers worry that AI could be expanded to allow it to deceive us. Meta’s CICERO AI proves how deceptive AI can be and how skillfully it can lie.
The world of artificial intelligence continues to evolve rapidly. We’re seeing the success of productive AI chatbots like ChatGPT, and many companies are working to incorporate AI into their applications and programs. Meanwhile, while the threat of AI remains a big issue, researchers have raised some interesting concerns about how easily AI lies to us and what this could mean going forward.
One of the things that complicates the use of ChatGPT and other AI systems is their tendency to make up information on the fly, so-called “hallucinations”. This is a shortcoming in the way AI works, and researchers worry that it could be extended to allow AI to fool us even more.
But can artificial intelligence really lie to us? It’s an interesting question, and some researchers believe they can answer it. According to the researchers, Meta’s CICERO AI is one of the most disturbing examples of how deceptive AI can be. This model was designed to play Diplomacy, and Meta says it was designed to be “largely honest and helpful”.

But looking at the data from the CICERO experiment, the researchers say that CICERO is a lie detector. CICERO even went so far as to work with a human player to pre-plan to trick another human player into leaving itself open to invasion.
He did this by conspiring with the German player and then working with the British player to get them to leave an opening in the North Sea. It did this by conspiring with Germany‘s player and then working with England’s player to get them to leave an opening in the North Sea. Above you can see the evidence of how AI lies and works against players to trick them into success. This is an interesting piece of evidence and just one of many examples researchers have recorded from the CICERO AI.
The risk is that it can be misused in several different ways. The researchers note in their report that the potential risk is “limited only by the imagination and technical know-how of the malicious actors”. It will be interesting to see where this behavior could lead in the future, especially if learning this behavior does not require a clear intent to deceive.
The post A new danger: Artificial intelligence has also learned how to cheat humans appeared first on NSF - Magazine.