AI designed to communicate and cooperate in Diplomacy board game

Publicly released:
International

Researchers at DeepMind, a lab owned by Google's company, have developed an artificial intelligence (AI) agent that can negotiate and form agreements, allowing it to play the board game Diplomacy with other artificial agents. When developing AI, cooperating and communicating between other agents is important, and researchers have found that the board game Diplomacy is a useful test bed for this kind of behaviour, as it involves involves complex communication, negotiation and alliance formation between the players. The team designed a deep reinforcement learning approach that allowed agents to agree future moves and identify beneficial deals by predicting possible future game states. The team hopes the findings from the study can help develop AI that can adapt their strategies to the environment.

Media release

From: Springer Nature

Artificial intelligence: Designing agents that can communicate and cooperate in Diplomacy

Artificial intelligence (AI) agents that can negotiate and form agreements, allowing them to outperform other agents without this ability in the board game Diplomacy, are reported in a Nature Communications paper. The findings demonstrate a deep reinforcement learning approach for modelling agents that can communicate and cooperate with other artificial agents to make joint plans when playing the game.

Developing AI that can demonstrate cooperation and communication between agents is important. Diplomacy is a popular board game that offers a useful test bed for such behaviour, as it involves complex communication, negotiation and alliance formation between the players, which have been long-lasting challenges for AI to achieve. To play successfully, Diplomacy requires reasoning about concurrent player future plans, commitments between players and their honest cooperation. Previous AI agents have achieved success in single-player or competitive two-player games without communication between players.

János Kramár, Yoram Bachrach and colleagues designed a deep reinforcement learning approach that enables agents to negotiate alliances and joint plans. The authors created agents that model game players and form teams that try to counter the strategies of other teams. The learning algorithm allows agents to agree future moves and identify beneficial deals by predicting possible future game states. Moving towards human-level performance, the authors also investigated the conditions for honest cooperation, by examining some broken commitment scenarios between the agents, where agents deviate from past agreements.

The findings help form the basis of flexible communication mechanisms in AI agents that enable them to adapt their strategies to their environment. Additionally, the findings show how the inclination to sanction peers who break contracts dramatically reduces the advantage of such deviators, and helps foster mostly truthful communication, despite conditions that initially favour deviations from agreements.

Attachments

Note: Not all attachments are visible to the general public. Research URLs will go live after the embargo ends.

Research Springer Nature, Web page The URL will go live after the embargo lifts.
Journal/
conference:
Nature Communications
Research:Paper
Organisation/s: DeepMind, UK
Funder: N/A; The authors declare no competing interests.
Media Contact/s
Contact details are only visible to registered journalists.