Update README.md
Browse files
README.md
CHANGED
|
@@ -57,7 +57,7 @@ pretty_name: U3T Monte Carlo Tree Search Position Evaluations
|
|
| 57 |
size_categories:
|
| 58 |
- 1M<n<10M
|
| 59 |
---
|
| 60 |
-
# Dataset Card for
|
| 61 |
|
| 62 |
This dataset stores refactored data of [Monte Carlo Search Tree Evaluations](https://en.wikipedia.org/wiki/Monte_Carlo_tree_search) edited from arnowaczynski's [utttai](https://github.com/arnowaczynski/utttai), for the game [Ultimate-tic-tac-toe](https://en.wikipedia.org/wiki/Ultimate_tic-tac-toe) (U3T).
|
| 63 |
One can use this data to train a model to evaluate positions or predict moves.
|
|
@@ -67,12 +67,12 @@ One can use this data to train a model to evaluate positions or predict moves.
|
|
| 67 |
This dataset contains over 8 million evaluated positions at varying depth (via [utttai](https://github.com/arnowaczynski/utttai/tree/main/datasets)):
|
| 68 |

|
| 69 |
|
| 70 |
-
Each evaluated positions counts the number of wins, draws, and loses, giving an estimation as to how good a position is. Each position stores an array of moves, and their respective
|
| 71 |
MCTS searches.
|
| 72 |
|
| 73 |
## Uses
|
| 74 |
|
| 75 |
-
One can use this dataset in an way they want. But, it is mainly intended for making a deep learning model to play U3T.
|
| 76 |
|
| 77 |
## Dataset Structure
|
| 78 |
|
|
@@ -118,7 +118,10 @@ This dataset stores evaluations for game positions in Ultimate Tic-Tac-Toe. Each
|
|
| 118 |
|
| 119 |
## Dataset Creation
|
| 120 |
|
| 121 |
-
Documentation on the generation of this dataset can be found [here](https://github.com/markstanl/uttt-bots/tree/main/data/hugging_face).
|
|
|
|
|
|
|
|
|
|
| 122 |
|
| 123 |
### Dataset Splitting
|
| 124 |
|
|
@@ -133,10 +136,10 @@ The dataset was split into a train, test, and validation set, with a distributio
|
|
| 133 |
### Curation Rationale
|
| 134 |
|
| 135 |
The original dataset by arnowaczynski was not stored in a conventional big data format, and used some unconventional index storing [(detailed here)](https://github.com/markstanl/uttt-bots/blob/main/utttai_conversion/utttai.md).
|
| 136 |
-
So, we decided to refactor it and upload it here, to Hugging Face.
|
| 137 |
|
| 138 |
|
| 139 |
### Source Data
|
| 140 |
|
| 141 |
-
arnowaczynski generated these datasets in their [GitHub Repository](https://github.com/arnowaczynski/utttai/tree/main/datasets). Further documentation on how
|
| 142 |
the dataset was generated is available at that link.
|
|
|
|
| 57 |
size_categories:
|
| 58 |
- 1M<n<10M
|
| 59 |
---
|
| 60 |
+
# Dataset Card for U3T
|
| 61 |
|
| 62 |
This dataset stores refactored data of [Monte Carlo Search Tree Evaluations](https://en.wikipedia.org/wiki/Monte_Carlo_tree_search) edited from arnowaczynski's [utttai](https://github.com/arnowaczynski/utttai), for the game [Ultimate-tic-tac-toe](https://en.wikipedia.org/wiki/Ultimate_tic-tac-toe) (U3T).
|
| 63 |
One can use this data to train a model to evaluate positions or predict moves.
|
|
|
|
| 67 |
This dataset contains over 8 million evaluated positions at varying depth (via [utttai](https://github.com/arnowaczynski/utttai/tree/main/datasets)):
|
| 68 |

|
| 69 |
|
| 70 |
+
Each evaluated positions counts the number of wins, draws, and loses, giving an estimation as to how good a position is. Each position stores an array of legal moves, and their respective
|
| 71 |
MCTS searches.
|
| 72 |
|
| 73 |
## Uses
|
| 74 |
|
| 75 |
+
One can use this dataset in an way they want. But, it is mainly intended for making a deep learning model to play U3T, or evaluate a static position.
|
| 76 |
|
| 77 |
## Dataset Structure
|
| 78 |
|
|
|
|
| 118 |
|
| 119 |
## Dataset Creation
|
| 120 |
|
| 121 |
+
Documentation on the generation of this dataset can be found [here](https://github.com/markstanl/uttt-bots/tree/main/data/hugging_face).
|
| 122 |
+
Roughly, we converted the UTTTAI gamestate to the standardized version documented before. Processed the txt file into JSONL files for each depth.
|
| 123 |
+
Converted each JSON line in the JSONL to a dictionary with the new key of depth. Appended a list with all of the dictionaries.
|
| 124 |
+
Stored it as a large parquet file. Then split the data accordingly.
|
| 125 |
|
| 126 |
### Dataset Splitting
|
| 127 |
|
|
|
|
| 136 |
### Curation Rationale
|
| 137 |
|
| 138 |
The original dataset by arnowaczynski was not stored in a conventional big data format, and used some unconventional index storing [(detailed here)](https://github.com/markstanl/uttt-bots/blob/main/utttai_conversion/utttai.md).
|
| 139 |
+
So, we decided to refactor it and upload it here, to Hugging Face, in a dataset that is more efficient for NN training.
|
| 140 |
|
| 141 |
|
| 142 |
### Source Data
|
| 143 |
|
| 144 |
+
arnowaczynski generated these datasets in their [GitHub Repository](https://github.com/arnowaczynski/utttai/tree/main/datasets). Further documentation on how
|
| 145 |
the dataset was generated is available at that link.
|