Cicero - The AI That Won a Game of Betrayal... by Being Nice???
What Meta’s "Diplomacy" Bot Teaches Us About the Future of Trust
by Dan Roque | Reading Time: 6 mins | In Bots of the Future
Why We’re Looking at This Now
The current AI landscape is filled with irrelevant noise, unearned hype, and extreme doomsaying; for professionals, the challenge
isn't finding more data—it’s cutting through the exhausting cycle of
"doom-and-hype" to find the strategic signal. If we want to move past
seeing AI as a magic trick and start viewing it as a tool for high-stakes
collaboration, we have to look at the mechanisms, not just the headlines.
That brings us to CICERO, a breakthrough from Meta
AI. CICERO is the first artificial intelligence to achieve human-level
performance in the game Diplomacy. Now, if you haven’t played, don’t
think of this like Chess. There are no fixed patterns or isolated moves. Diplomacy
is a game of people. It’s a seven-player mix of Risk, Poker, and the TV
show Survivor. There are no dice and no luck; there is only negotiation.
To win, you must form alliances, coordinate simultaneous moves, and—most
importantly—keep or break trust.
Today, we’re going to step away from the "AI Wins
Game" scoreboard and pick up the chalk. We need to understand not just what
CICERO did, but the specific architecture of how it actually thinks.
A Look Under the Hood: The Strategic Architecture
If I were standing at a chalkboard right now, I’d draw two
large boxes with a thick, double-sided arrow running between them. CICERO’s
breakthrough isn't just "better chat"; it is the integration of two
distinct AI disciplines that have historically lived in different worlds.
- Box
A: The Planning Engine (Strategic Reasoning): This is the
"Tactician." It looks at the board state and the history of the
game to model what the other six human players are likely to do. It isn't
just looking for the "best move" in a vacuum; it’s predicting
human intent.
- Box
B: The Dialogue Model (Natural Language Processing): This is the
"Spokesperson." It takes the tactical goals from the Planning
Engine and translates them into free-form text.
The Breakthrough: Grounded Planning
In the early days
of development, researchers noticed the bot had a great "arm" but no
"head"—it was tactically strong but lacked a strategic sense, often
attacking for short-term gains while burning long-term bridges. The fix? The Planning
Engine began "controlling" the Dialogue Model.
This is what we call Grounded Planning. The words
aren't just floating in space or predicting the "next likely word"
like a standard chatbot. They are tethered to a physical move on the board. The
message is a tool to execute a specific plan. This is a massive leap: it’s the
difference between a bot that can talk about a plan and a bot that can
actually execute one with you.
Honesty is a Strategic Hack
In a game legendary for its "stabs" and betrayals,
the most shocking finding from the research was that CICERO was almost
entirely honest. In a zero-sum environment with no binding agreements,
honesty became the ultimate competitive advantage.
The "Grudge" Mechanism Why not lie? Because
lying is computationally and socially expensive. Humans hold grudges. In a
three-hour game of Diplomacy, once a lie is detected, there is often no
going back. The researchers found that while a human might try to navigate the
messy "emotional fallout" of a betrayal, the bot defaults to honesty
because it cannot effectively model the "grudge" a human player will
maintain for the next ten turns. Honesty wasn't a moral choice; it was a way to
prevent the permanent break of social capital.
Perception vs. Reality:
- The
Filtering Shield: To prevent the bot from being exploited or
accidentally leaking its own moves while being "too honest,"
Meta used a Message Filtering Mechanism to block messages that
could be used against the bot.
- The
Liar’s Paradox: Interestingly, humans felt the bot lied more
than it did. In reality, human players broke commitments 1.2%–1.5% of the
time, while the bot stayed below 1%.
By offering helpful advice and staying reliable, the bot
built stronger alliances than the humans who were busy trying to out-trick one
another.
The "France Factor" and the Limits of Persuasion
Despite the bot's success, the Wongkamjan et al. study
(2024) gives us a vital reality check: position matters more than talk.
The 14x Multiplier The study identified the "France
Factor." In Diplomacy, France is an inherently strong power.
The data showed that starting as a strong power had a 14x larger impact
on success than the specific communication strategy used.
However, there’s a strategic signal in the messy middle:
- Austria/Turkey:
CICERO dominates here because it makes better tactical decisions in
clear-cut environments.
- Italy:
This is where humans held their ground. Italy requires "careful
coordination of actions" and managing messy, multi-front
relationships. In these high-coordination, "messy-middle"
scenarios, humans are still comparable to the machine.
The Persuasion Gap We also see the limits of
"Transactional" AI. Humans can eventually spot the bot because its
talk is move-oriented—it lacks the "How are you today?" emotional
glue of top-tier human players. This reflected in the Persuasion Success
Rates:
- Human-to-Human
Persuasion: 21.1% success (Humans convincing humans).
- CICERO-to-Human
Persuasion: 10.9% success (The bot convincing a human).
- Human-to-CICERO Persuasion: 8.6% success (The bot is "stubborn" because it is grounded in its Planning Engine, making it less susceptible to emotional appeals).
The Future of Collaborative AI
If you take one thing from the chalkboard today, let it be
this: AI is a tool to be mastered.
The CICERO research helps us solve the "hard
problems" of the next decade: grounded planning, understanding
reciprocity, and navigating deception. These "silly games" are the
high-altitude training grounds for the real world.
When we see a bot successfully allying with a world
champion, we aren't just looking at a game; we are looking at the precursor to
AI that can help us coordinate disaster relief, manage global supply chains, or
negotiate complex contracts. The future isn't a machine that just
"talks"—it’s a machine that can collaborate with us to solve problems
in a complex, human world.
Works Cited
AI at Meta. “CICERO.” AI at Meta, Meta Platforms, Inc., n.d., https://ai.meta.com/research/cicero/. Accessed 24 Mar. 2026.
AI at Meta. “Diplomacy and Meta AI’s CICERO.” AI at Meta, Meta Platforms, Inc., n.d., https://ai.meta.com/research/cicero/diplomacy/. Accessed 24 Mar. 2026.
“Artificial Intelligence DEFEATS Humans At DIPLOMACY …” YouTube, n.d., https://www.youtube.com/watch?v=Y8bEKCGNA04. Accessed 24 Mar. 2026.
Bakhtin, Anton, et al. “Human-level Play in the Game of Diplomacy by Combining Language Models with Strategic Reasoning.” Science, vol. 378, no. 6624, 2 Dec. 2022, pp. 1067–1074. https://doi.org/10.1126/science.ade9097. Accessed 25 Mar. 2026.
CaptainMeme. “Expert Diplomacy Player vs CICERO AI.” YouTube, n.d., https://www.youtube.com/watch?v=u5192bvUS7k. Accessed 25 Mar. 2026.
Edwards, Benj. “Meta Researchers Create AI That Masters Diplomacy, Tricking Human Players.” Ars Technica, 22 Nov. 2022, https://arstechnica.com/information-technology/2022/11/meta-researchers-create-ai-that-masters-diplomacy-tricking-human-players/. Accessed 26 Mar. 2026.
facebookresearch. “diplomacy_cicero.” GitHub, Meta Platforms, Inc., 17 Apr. 2025, https://github.com/facebookresearch/diplomacy_cicero. Accessed 26 Mar. 2026.
“[R] Human-level Play in the Game of Diplomacy by Combining Language Models with Strategic Reasoning — Meta AI.” Reddit, r/MachineLearning, n.d., https://www.reddit.com/r/MachineLearning/comments/z1yt45/r_humanlevel_play_in_the_game_of_diplomacy_by/. Accessed 26 Mar. 2026.
Meta AI. “CICERO: An AI Agent That Negotiates, Persuades, and Cooperates with People.” AI at Meta, Meta Platforms, Inc., 22 Nov. 2022, https://ai.meta.com/blog/cicero-ai-negotiates-persuades-and-cooperates-with-people/. Accessed 27 Mar. 2026.
“Meta Released a Diplomacy-playing LLM. How Good Is Cicero at Talking to Players? [Research Talk].” YouTube, n.d., https://www.youtube.com/watch?v=fKaiumkZjL4. Accessed 27 Mar. 2026.
Walsh, Toby. “An AI Named Cicero Can Beat Humans in Diplomacy, a Complex Alliance-building Game. Here’s Why That’s a Big Deal.” The Conversation, 24 Nov. 2022, https://theconversation.com/an-ai-named-cicero-can-beat-humans-in-diplomacy-a-complex-alliance-building-game-heres-why-thats-a-big-deal-195208. Accessed 27 Mar. 2026.
Wongkamjan, Wichayaporn, et al. “More Victories, Less Cooperation: Assessing Cicero’s Diplomacy Play.” arXiv, 7 June 2024, arXiv:2406.04643, https://arxiv.org/abs/2406.04643. Accessed 28 Mar. 2026.

Comments
Post a Comment