Cicero - The AI That Won a Game of Betrayal... by Being Nice???

What Meta’s "Diplomacy" Bot Teaches Us About the Future of Trust

by Dan Roque | Reading Time: 6 mins | In Bots of the Future

Why We’re Looking at This Now

The current AI landscape is filled with irrelevant noise, unearned hype, and extreme doomsaying; for professionals, the challenge isn't finding more data—it’s cutting through the exhausting cycle of "doom-and-hype" to find the strategic signal. If we want to move past seeing AI as a magic trick and start viewing it as a tool for high-stakes collaboration, we have to look at the mechanisms, not just the headlines.

That brings us to CICERO, a breakthrough from Meta AI. CICERO is the first artificial intelligence to achieve human-level performance in the game Diplomacy. Now, if you haven’t played, don’t think of this like Chess. There are no fixed patterns or isolated moves. Diplomacy is a game of people. It’s a seven-player mix of Risk, Poker, and the TV show Survivor. There are no dice and no luck; there is only negotiation. To win, you must form alliances, coordinate simultaneous moves, and—most importantly—keep or break trust.

Today, we’re going to step away from the "AI Wins Game" scoreboard and pick up the chalk. We need to understand not just what CICERO did, but the specific architecture of how it actually thinks.


A Look Under the Hood: The Strategic Architecture

If I were standing at a chalkboard right now, I’d draw two large boxes with a thick, double-sided arrow running between them. CICERO’s breakthrough isn't just "better chat"; it is the integration of two distinct AI disciplines that have historically lived in different worlds.

  • Box A: The Planning Engine (Strategic Reasoning): This is the "Tactician." It looks at the board state and the history of the game to model what the other six human players are likely to do. It isn't just looking for the "best move" in a vacuum; it’s predicting human intent.
  • Box B: The Dialogue Model (Natural Language Processing): This is the "Spokesperson." It takes the tactical goals from the Planning Engine and translates them into free-form text.

The Breakthrough: Grounded Planning

In the early days of development, researchers noticed the bot had a great "arm" but no "head"—it was tactically strong but lacked a strategic sense, often attacking for short-term gains while burning long-term bridges. The fix? The Planning Engine began "controlling" the Dialogue Model.

This is what we call Grounded Planning. The words aren't just floating in space or predicting the "next likely word" like a standard chatbot. They are tethered to a physical move on the board. The message is a tool to execute a specific plan. This is a massive leap: it’s the difference between a bot that can talk about a plan and a bot that can actually execute one with you. 


Honesty is a Strategic Hack

In a game legendary for its "stabs" and betrayals, the most shocking finding from the research was that CICERO was almost entirely honest. In a zero-sum environment with no binding agreements, honesty became the ultimate competitive advantage.

The "Grudge" Mechanism Why not lie? Because lying is computationally and socially expensive. Humans hold grudges. In a three-hour game of Diplomacy, once a lie is detected, there is often no going back. The researchers found that while a human might try to navigate the messy "emotional fallout" of a betrayal, the bot defaults to honesty because it cannot effectively model the "grudge" a human player will maintain for the next ten turns. Honesty wasn't a moral choice; it was a way to prevent the permanent break of social capital.

Perception vs. Reality:

  • The Filtering Shield: To prevent the bot from being exploited or accidentally leaking its own moves while being "too honest," Meta used a Message Filtering Mechanism to block messages that could be used against the bot.
  • The Liar’s Paradox: Interestingly, humans felt the bot lied more than it did. In reality, human players broke commitments 1.2%–1.5% of the time, while the bot stayed below 1%.

By offering helpful advice and staying reliable, the bot built stronger alliances than the humans who were busy trying to out-trick one another.


The "France Factor" and the Limits of Persuasion

Despite the bot's success, the Wongkamjan et al. study (2024) gives us a vital reality check: position matters more than talk.

The 14x Multiplier The study identified the "France Factor." In Diplomacy, France is an inherently strong power. The data showed that starting as a strong power had a 14x larger impact on success than the specific communication strategy used.

However, there’s a strategic signal in the messy middle:

  • Austria/Turkey: CICERO dominates here because it makes better tactical decisions in clear-cut environments.
  • Italy: This is where humans held their ground. Italy requires "careful coordination of actions" and managing messy, multi-front relationships. In these high-coordination, "messy-middle" scenarios, humans are still comparable to the machine.

The Persuasion Gap We also see the limits of "Transactional" AI. Humans can eventually spot the bot because its talk is move-oriented—it lacks the "How are you today?" emotional glue of top-tier human players. This reflected in the Persuasion Success Rates:

  • Human-to-Human Persuasion: 21.1% success (Humans convincing humans).
  • CICERO-to-Human Persuasion: 10.9% success (The bot convincing a human).
  • Human-to-CICERO Persuasion: 8.6% success (The bot is "stubborn" because it is grounded in its Planning Engine, making it less susceptible to emotional appeals).

The Future of Collaborative AI

If you take one thing from the chalkboard today, let it be this: AI is a tool to be mastered.

The CICERO research helps us solve the "hard problems" of the next decade: grounded planning, understanding reciprocity, and navigating deception. These "silly games" are the high-altitude training grounds for the real world.

When we see a bot successfully allying with a world champion, we aren't just looking at a game; we are looking at the precursor to AI that can help us coordinate disaster relief, manage global supply chains, or negotiate complex contracts. The future isn't a machine that just "talks"—it’s a machine that can collaborate with us to solve problems in a complex, human world.


Works Cited

AI at Meta. “CICERO.” AI at Meta, Meta Platforms, Inc., n.d., https://ai.meta.com/research/cicero/. Accessed 24 Mar. 2026.

AI at Meta. “Diplomacy and Meta AI’s CICERO.” AI at Meta, Meta Platforms, Inc., n.d., https://ai.meta.com/research/cicero/diplomacy/. Accessed 24 Mar. 2026.

“Artificial Intelligence DEFEATS Humans At DIPLOMACY …” YouTube, n.d., https://www.youtube.com/watch?v=Y8bEKCGNA04Accessed 24 Mar. 2026.

Bakhtin, Anton, et al. “Human-level Play in the Game of Diplomacy by Combining Language Models with Strategic Reasoning.” Science, vol. 378, no. 6624, 2 Dec. 2022, pp. 1067–1074. https://doi.org/10.1126/science.ade9097. Accessed 25 Mar. 2026.

CaptainMeme. “Expert Diplomacy Player vs CICERO AI.” YouTube, n.d., https://www.youtube.com/watch?v=u5192bvUS7k. Accessed 25 Mar. 2026.

Edwards, Benj. “Meta Researchers Create AI That Masters Diplomacy, Tricking Human Players.”  Ars Technica, 22 Nov. 2022, https://arstechnica.com/information-technology/2022/11/meta-researchers-create-ai-that-masters-diplomacy-tricking-human-players/. Accessed 26 Mar. 2026.

facebookresearch. “diplomacy_cicero.” GitHub, Meta Platforms, Inc., 17 Apr. 2025, https://github.com/facebookresearch/diplomacy_cicero. Accessed 26 Mar. 2026.

“[R] Human-level Play in the Game of Diplomacy by Combining Language Models with Strategic Reasoning — Meta AI.” Reddit, r/MachineLearning, n.d., https://www.reddit.com/r/MachineLearning/comments/z1yt45/r_humanlevel_play_in_the_game_of_diplomacy_by/. Accessed 26 Mar. 2026.

Meta AI. “CICERO: An AI Agent That Negotiates, Persuades, and Cooperates with People.” AI at Meta, Meta Platforms, Inc., 22 Nov. 2022, https://ai.meta.com/blog/cicero-ai-negotiates-persuades-and-cooperates-with-people/. Accessed 27 Mar. 2026.

“Meta Released a Diplomacy-playing LLM. How Good Is Cicero at Talking to Players? [Research Talk].” YouTube, n.d., https://www.youtube.com/watch?v=fKaiumkZjL4Accessed 27 Mar. 2026.

Walsh, Toby. “An AI Named Cicero Can Beat Humans in Diplomacy, a Complex Alliance-building Game. Here’s Why That’s a Big Deal.” The Conversation, 24 Nov. 2022, https://theconversation.com/an-ai-named-cicero-can-beat-humans-in-diplomacy-a-complex-alliance-building-game-heres-why-thats-a-big-deal-195208. Accessed 27 Mar. 2026.

Wongkamjan, Wichayaporn, et al. “More Victories, Less Cooperation: Assessing Cicero’s Diplomacy Play.” arXiv, 7 June 2024, arXiv:2406.04643, https://arxiv.org/abs/2406.04643. Accessed 28 Mar. 2026.


Watch and/or Listen

Youtube | Spotify Apple Podcasts



Comments

Read More CasiornThinks Takes

Smallville: The Secret Lives of Generative Agents

AI and Your Career: Navigating Change with Confidence (HRUCKUS Feature)

Our Season 1 Roadmap Is Here!