Have the highest score at the end of the game.

Generated at 20/02/2021, 01:13 from 1000 logged games.

Representative game (in the sense of being of mean length). Wherever you see the 'representative game' referred to in later sections, this is it!

Each turn, either:

- Place a stone of your colour in an empty space so as to create a new group.
- Grow all of your groups on the board.

When the board is full, you score as follows:*{Stones placed} - k x {number of groups}*

Where k is a multiple of 4 chosen at the start of the game.

General comments:

Play: Combinatorial

Mechanism(s): Connection,Pattern

Components: Board

BGG Entry | HexSymple(19x19, group penalty=12) |
---|---|

BGG Rating | 7.97 |

#Voters | 10 |

SD | 0.5728 |

BGG Weight | 0 |

#Voters | 0 |

Year | 2010 |

User | Rating | Comment |
---|---|---|

alekerickson | 9 | innovative! |

luigi87 | 8.2 | |

mrraow | 8 | Clever territory game; the branching factor is brutal, so don't expect a strong AI! |

T0afer | 8 | |

RichardV | 8.5 | Great abstract game that distils something of the essence of Go and Hex with very simple rules. There are several distinct phases to the game, from early proliferation, to later block expansion, and then a final push to consolidate groups. The java ai on ai ai is pretty good to get started. |

kevan | 7 | |

milomilo122 | N/A | A beautiful idea that doesn't quite work as a game for me, because too many stones need to be placed on a single turn. When I have to place 8 stones on a turn in a deep game like this, it's a recipe for titanic analysis paralysis. |

orangeblood | 8 | This is a very interesting design with mechanics that I've not seen before. You could call it a Go variant, but I think it’s less like Go than Blooms, for example (where territory and captures matter). As concisely as I can explain, on your turn you either place a single stone to start a new group, or add one stone to each of your existing groups. When the board is filled you score one point per stone, minus a predetermined number (e.g., 6 points) for every group you have. If you’re somewhat dense like me it will take you a few of plays to begin to understand the delicate early-game balance between starting new groups, vs. switching gears and adding to existing ones. The math tells you that the more groups you have the more stones you’ll eventually be able to add… and yet the higher group penalty points at the end. That brings up the entire strategy of positioning your stones — where Go-like play is generally rewarded. For instance, it’s nice if you could join up groups toward the end to avoid penalty points, or deny your opponent the same…. and to have staked out some territory to give your groups room to grow. Co-designed by Christian Freeling and Benedikt Rosenau, Christian says it’s one of only six games of his that he considers relevant. Or, as he also says, in as much as abstract games matter, this one matters. |

simpledeep | 8 | |

Zapawa | 8 | This game should have been created thousands of years ago and been played as a classic ever since. It's versatile, simple, sharp and very original. Sadly, it's only 10 years old and quite obscure at this point, but I hope it's time of glory comes within my lifetime. For all my admiration for the idea, I enjoyed the reality of it far less than I thought I would -- I'm just too dumb to appreciate the strategies here. Might increase the rating if I ever learn to play it properly. |

rchandra | 7 |

Size (bytes) | 33297 |
---|---|

Reference Size | 10293 |

Ratio | 3.23 |

Ai Ai calculates the size of the implementation, and compares it to the Ai Ai implementation of the simplest possible game (which just fills the board). Note that this estimate may include some graphics and heuristics code as well as the game logic. See the wikipedia entry for more details.

Playouts per second | 1853.14 (539.62µs/playout) |
---|---|

Reference Size | 372286.96 (2.69µs/playout) |

Ratio (low is good) | 200.90 |

Tavener complexity: the heat generated by playing every possible instance of a game with a perfectly efficient programme. Since this is not possible to calculate, Ai Ai calculates the number of random playouts per second and compares it to the fastest non-trivial Ai Ai game (Connect 4). This ratio gives a practical indication of how complex the game is. Combine this with the computational state space, and you can get an idea of how strong the default (MCTS-based) AI will be.

Label | Its/s | SD | Nodes/s | SD | Game length | SD |
---|---|---|---|---|---|---|

Random playout | 1,939 | 16 | 525,536 | 4,457 | 271 | 0 |

search.UCB | 1,856 | 56 | 271 | 0 | ||

search.UCT | 1,880 | 18 | 271 | 0 |

Random: 10 second warmup for the hotspot compiler. 100 trials of 1000ms each.

Other: 100 playouts, means calculated over the first 5 moves only to avoid distortion due to speedup at end of game.

Rotation (Half turn) lost each game as expected.

Reflection (X axis) lost each game as expected.

Reflection (Y axis) lost each game as expected.

Copy last move lost each game as expected.

Mirroring strategies attempt to copy the previous move. On first move, they will attempt to play in the centre. If neither of these are possible, they will pick a random move. Each entry represents a different form of copying; direct copy, reflection in either the X or Y axis, half-turn rotation.

1: White win % | 48.10±3.08 | Includes draws = 50% |
---|---|---|

2: Black win % | 51.90±3.10 | Includes draws = 50% |

Draw % | 0.00 | Percentage of games where all players draw. |

Decisive % | 100.00 | Percentage of games with a single winner. |

Samples | 1000 | Quantity of logged games played |

Note: that win/loss statistics may vary depending on thinking time (horizon effect, etc.), bad heuristics, bugs, and other factors, so should be taken with a pinch of salt. (Given perfect play, any game of pure skill will always end in the same result.)

Note: Ai Ai differentiates between states where all players draw or win or lose; this is mostly to support cooperative games.

Match | AI | Strong Wins | Draws | Strong Losses | #Games | Strong Score | p1 Win% | Draw% | p2 Win% | Game Length |
---|---|---|---|---|---|---|---|---|---|---|

0 | Random | |||||||||

1 | UCT (its=2) | 631 | 0 | 345 | 976 | 0.6160 <= 0.6465 <= 0.6759 | 51.74 | 0.00 | 48.26 | 271.00 |

6 | UCT (its=7) | 631 | 0 | 338 | 969 | 0.6206 <= 0.6512 <= 0.6805 | 50.15 | 0.00 | 49.85 | 271.00 |

10 | UCT (its=11) | 575 | 0 | 425 | 1000 | 0.5441 <= 0.5750 <= 0.6053 | 49.30 | 0.00 | 50.70 | 271.00 |

11 | UCT (its=11) | 504 | 0 | 496 | 1000 | 0.4731 <= 0.5040 <= 0.5349 | 49.80 | 0.00 | 50.20 | 271.00 |

Search for levels ended: time limit reached.

Level of Play: **Strong** beats **Weak** 60% of the time (lower bound with 95% confidence).

Draw%, p1 win% and game length may give some indication of trends as AI strength increases.

This chart shows the win(green)/draw(black)/loss(red) percentages, as UCT play strength increases. **Note that for most games, the top playing strength show here will be distinctly below human standard.**

Game length | 271.01 | |
---|---|---|

Branching factor | 59.02 | |

Complexity | 10^394.63 | Based on game length and branching factor |

Samples | 1000 | Quantity of logged games played |

Computational complexity (where present) is an estimate of the game tree reachable through actual play. For each game in turn, Ai Ai marks the positions reached in a hashtable, then counts the number of new moves added to the table. Once all moves are applied, it treats this sequence as a geometric progression and calculates the sum as n-> infinity.

Distinct actions | 272 | Number of distinct moves (e.g. "e4") regardless of position in game tree |
---|---|---|

Good moves | 82 | A good move is selected by the AI more than the average |

Bad moves | 190 | A bad move is selected by the AI less than the average |

Response distance | 9.08 | Mean distance between move and response; a low value relative to the board size may mean a game is tactical rather than strategic. |

Samples | 1000 | Quantity of logged games played |

A mean of 100.00% of board locations were used per game.

Colour and size show the frequency of visits.

Game length frequencies.

Mean | 271.01 |
---|---|

Mode | [271] |

Median | 271.0 |

This chart is based on a single representative* playout, and gives a feel for the change in material over the course of a game. (* Representative in the sense that it is close to the mean length.)

Table: branching factor per turn, based on a single representative* game. (* Representative in the sense that it is close to the mean game length.)

This chart is based on a single representative* game, and gives a feel for the types of moves available throughout that game. (* Representative in the sense that it is close to the mean game length.)

Red: removal, Black: move, Blue: Add, Grey: pass, Purple: swap sides, Brown: other.

Colour shows the success ratio of this play over the first 10moves; black < red < yellow < white.

Size shows the frequency this move is played.

0 | 1 | 2 |
---|---|---|

1 | 271 | 73441 |

Note: most games do not take board rotation and reflection into consideration.

Multi-part turns could be treated as the same or different depth depending on the implementation.

Counts to depth N include all moves reachable at lower depths.

Inaccuracies may also exist due to hash collisions, but Ai Ai uses 64-bit hashes so these will be a very small fraction of a percentage point.

No solutions found to depth 2.