This environment is part of the classic environments. Please read that page first for general information.


from pettingzoo.classic import hanabi_v4



Parallel API


Manual Control



agents= ['player_0', 'player_1']



Action Shape


Action Values


Observation Shape


Observation Values


Hanabi is a 2-5 player cooperative game where players work together to form fireworks of different colors. A firework is a set of cards of the same color, ordered from 1 to 5. Cards in the game have both a color and number; each player can view the cards another player holds, but not their own. Players cannot directly communicate with each other, but must instead remove an info token from play in order to give information. Players can tell other players which of the cards in their hand is a specific color, or a specific number. There are initially 8 info tokens, but players can discard cards in their hand to return an info token into play. Players can also play a card from their hand: the card must either begin a new firework or be appended in order to an existing firework. However, 2 fireworks cannot have the same color, and a single firework cannot repeat numbers. If the played card does not satisfy these conditions, a life token is placed. The game ends when either 3 life tokens have been placed, all 5 fireworks have been completed, or all cards have been drawn from the deck. Points are awarded based on the largest card value in each created firework.

Environment arguments#

Hanabi takes in a number of arguments defining the size and complexity of the game. Default is a full 2 player hanabi game.

hanabi_v4.env(colors=5, rank=5, players=2, hand_size=5, max_information_tokens=8,
max_life_tokens=3, observation_type=1)

colors: Number of colors the cards can take (affects size of deck)

rank: Number of ranks the cards can take (affects size of deck)

hand_size: Size of player’s hands. Standard game is (4 if players >= 4 else 5)

max_information_tokens: Maximum number of information tokens (more tokens makes the game easier by allowing more information to be revealed)

max_life_tokens: Maximum number of life tokens (more tokens makes the game easier by allowing more information to be revealed)

observation_type: 0: Minimal observation. 1: First-order common knowledge observation (default).

Observation Space#

The observation is a dictionary which contains an 'observation' element which is the usual RL observation described below, and an 'action_mask' which holds the legal moves, described in the Legal Actions Mask section.

The main observation space of an agent is a 658 sized vector representing the life and info tokens left, the currently constructed fireworks, the hands of all other agents, the current deck size and the discarded cards. The observation vector contains the following features, life tokens, information tokens, number of players, deck size, formed fireworks, legal moves, observed hands, discard pile, the hints received from other players, which are then serialized into a bit string.

Each card is encoded with a 25 bit one-hot vector, where the encoding of a card is equal to its color*T + rank, where T is the max possible rank. By default this value is 5. The maximum deck size is 50. The remaining deck size is represented with unary encoding. The state of each colored firework is represented with a one-hot encoding. The information tokens remaining are represented with a unary encoding. The life tokens remaining are represented with a unary encoding. The discard pile is represented with a thermometer encoding of the ranks of each discarded card. That is the least significant bit being set to 1 indicates the lowest rank card of that color has been discarded.

As players reveal info about their cards, the information revealed per card is also observed. The first 25 bits represent whether or not that specific card could be a specific color. For example if the card could only be red, then the first 25 bits of the observed revealed info would be 11111 followed by 20 zeros. The next 5 bits store whether the color of that card was explicitly revealed, so if the card was revealed to be red, then the next 5 bits would be 10000. Finally the last 5 bits are the revealed rank of the card. So if the card was revealed to be of rank 1, then the next 5 bits would be 10000. These 25 bits are tracked and observed for all cards in each player’s hand.




0 - 24

Vector of Card 1 in other player’s hand

[0, 1]

25 - 49

Vector of Card 2 in other player’s hand

[0, 1]

50 - 74

Vector of Card 3 in other player’s hand

[0, 1]

75 -100

Vector of Card 4 in other player’s hand

[0, 1]


Vector of Card 5 in other player’s hand

[0, 1]


Unary Encoding of Remaining Deck Size

[0, 1]


Vector of Red Firework

[0, 1]


Vector of Yellow Firework

[0, 1]


Vector of Green Firework

[0, 1]


Vector of White Firework

[0, 1]


Vector of Blue Firework

[0, 1]


Unary Encoding of Remaining Info Tokens

[0, 1]


Unary Encoding of Remaining Life Tokens

[0, 1]


Thermometer Encoding of Discard Pile

[0, 1]


One-Hot Encoding of Previous Player ID

[0, 1]


Vector of Previous Player’s Action Type

[0, 1]


Vector of Target from Previous Action

[0, 1]


Vector of the Color Revealed in Last Action

[0, 1]


Vector of the Rank Revealed in Last Action

[0, 1]


Vector of Which Cards in the Hand were Revealed

[0, 1]


Position of the Card that was played or dropped

[0, 1]


Vector Representing Card that was last played

[0, 1]


Revealed Info of This Player’s 0th Card

[0, 1]


Revealed Info of This Player’s 1st Card

[0, 1]


Revealed Info of This Player’s 2nd Card

[0, 1]


Revealed Info of This Player’s 3rd Card

[0, 1]


Revealed Info of This Player’s 4th Card

[0, 1]


Revealed Info of Other Player’s 0th Card

[0, 1]


Revealed Info of Other Player’s 1st Card

[0, 1]


Revealed Info of Other Player’s 2nd Card

[0, 1]


Revealed Info of Other Player’s 3rd Card

[0, 1]


Revealed Info of Other Player’s 4th Card

[0, 1]

Action Space#

The action space is a scalar value, which ranges from 0 to the max number of actions. The values represent all possible actions a player can make, legal or not. Each possible move in the environment is mapped to a UUID, which ranges from 0 to the max number of moves. By default the max number of moves is 20. The first range of actions are to discard a card in the agent’s hand. If there are k cards in the player’s hand, then the first k action values are to discard one of those cards. The next k actions would be to play one of the cards in the player’s hand. Finally, the remaining actions are to reveal a color or rank in another players hand. The first set of reveal actions would be revealing all colors or values of cards for the next player in order, and this repeats for all the other players in the environment.

Action ID



Discard Card at position 0


Discard Card at position 1


Discard Card at position 2


Discard Card at position 3


Discard Card at position 4


Play Card at position 0


Play Card at position 1


Play Card at position 2


Play Card at position 3


Play Card at position 4


Reveal Red Cards for Player 1


Reveal Yellow Cards for Player 1


Reveal Green Cards for Player 1


Reveal White Cards for Player 1


Reveal Blue Cards for Player 1


Reveal Rank 1 Cards for Player 1


Reveal Rank 2 Cards for Player 1


Reveal Rank 3 Cards for Player 1


Reveal Rank 4 Cards for Player 1


Reveal Rank 5 Cards for Player 1


The reward of each step is calculated as the change in game score from the last step. The game score is calculated as the sum of values in each constructed firework. If the game is lost, the score is set to zero, so the final reward will be the negation of all reward received so far.

For example, if fireworks were created as follows:

Blue 1, Blue 2, Red 1, Green 1, Green 2, Green 3

At the end of the game, the total score would be 2 + 1 + 3 = 6

If an illegal action is taken, the game terminates and the one player that took the illegal action loses. Like an ordinary loss, their final reward will be the negation of all reward received so far. The reward of the other players will not be affected by the illegal action.

Version History#

  • v4: Fixed bug in arbitrary calls to observe() (1.8.0)

  • v3: Legal action mask in observation replaced illegal move list in infos (1.5.0)

  • v2: Fixed default parameters (1.4.2)

  • v1: Bumped version of all environments due to adoption of new agent iteration scheme where all agents are iterated over after they are done (1.4.0)

  • v0: Initial versions release (1.0.0)