Skip to content

Instantly share code, notes, and snippets.

@phimachine
Created December 2, 2019 00:13
Show Gist options
  • Save phimachine/a418ff601ef675023956cb22e89f3f7c to your computer and use it in GitHub Desktop.
Save phimachine/a418ff601ef675023956cb22e89f3f7c to your computer and use it in GitHub Desktop.
Current log
C:\Users\JasonHu\Anaconda3\envs\alphazero-checker\python.exe "C:\Users\JasonHu\AppData\Local\JetBrains\PyCharm 2018.3\helpers\pydev\pydevconsole.py" --mode=client --port=13753
import sys; print('Python %s on %s' % (sys.version, sys.platform))
sys.path.extend(['D:\\Git\\alphazero-checker', 'D:/Git/alphazero-checker'])
PyDev console: starting.
Python 3.6.8 |Anaconda, Inc.| (default, Dec 30 2018, 18:50:55) [MSC v.1915 64 bit (AMD64)] on win32
runfile('D:/Git/alphazero-checker/zero.py', wdir='D:/Git/alphazero-checker')
loading model at D:\Git\alphazero-checker\saves\lowpuct_7_0.pkl
Loaded model at epoch 7 iteration 0
Generating a new game with MCTS
Generating a new game with MCTS
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Game step 0 /200
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Game step 0 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 50 /200
Game step 40 /200
Game step 50 /200
Game step 50 /200
Game step 60 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 70 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 80 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 80 /200
Game step 80 /200
Game step 90 /200
Game step 80 /200
Game step 80 /200
Game step 80 /200
Game step 80 /200
Terminated due to peaceful activity
Terminated at step 82
First player won.
7[[ 2 -1 -1]
6 [ 2 -1 ]
5 [ 2 1 2 ]
4 [ 1 ]
3 [ 1]
2 [ -2 -2 ]
1 [ -1 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Terminated due to peaceful activity
Terminated at step 89
Second player won
7[[ -1]
6 [ 2 1 ]
5 [ 2 2 1]
4 [ 1 ]
3 [ -1 ]
2 [-1 -2 -2 ]
1 [ -1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 80 /200
Game step 100 /200
Game step 90 /200
Game step 90 /200
Game step 90 /200
Game step 10 /200
Game step 10 /200
Game step 90 /200
Game step 90 /200
Terminated due to peaceful activity
Terminated at step 94
First player won.
7[[ -1 -1]
6 [ 2 2 1 ]
5 [ 1 2 1]
4 [ -2 ]
3 [ ]
2 [ -2 ]
1 [ -1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 110 /200
Terminated due to peaceful activity
Terminated at step 111
Second player won
7[[ 2]
6 [ 2 1 ]
5 [ 1 2 ]
4 [ 2 ]
3 [ -2 ]
2 [-1 -2 ]
1 [ -1 -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 100 /200
Game step 20 /200
Game step 100 /200
Game step 100 /200
Game step 10 /200
Game step 10 /200
Game step 30 /200
Terminated due to peaceful activity
Terminated at step 107
Second player won
7[[ -1]
6 [ 2 1 ]
5 [ 2 1 2 ]
4 [ 1 ]
3 [ -2 ]
2 [-1 -2 ]
1 [ -1 -2 -2]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 100 /200
Game step 30 /200
Terminated due to peaceful activity
Terminated at step 104
First player won.
7[[ 2]
6 [ 2 ]
5 [ 1 2 ]
4 [ 2 1 ]
3 [ -1 ]
2 [ -2 -1 ]
1 [ 1 -2 -2]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 110 /200
Terminated due to peaceful activity
Terminated at step 112
First player won.
7[[ 2]
6 [ 2 1 ]
5 [ 2 2 1]
4 [-1 ]
3 [ -1 -1 -2 ]
2 [ -2 -1 ]
1 [ 1 -2 ]
0 [ 1 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 40 /200
Game step 10 /200
Game step 20 /200
Game step 10 /200
Terminated due to peaceful activity
Terminated at step 107
First player won.
7[[ 2 -1]
6 [ 2 -1 ]
5 [ 2 2 -1]
4 [ 1 -2 -1 ]
3 [ 2 -2 1]
2 [ -2 ]
1 [ ]
0 [ ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 40 /200
Game step 10 /200
Game step 30 /200
Game step 50 /200
Game step 20 /200
Game step 30 /200
Game step 20 /200
Game step 10 /200
Game step 20 /200
Game step 50 /200
Game step 40 /200
Game step 30 /200
Game step 60 /200
Game step 40 /200
Game step 20 /200
Game step 30 /200
Game step 30 /200
Game step 50 /200
Game step 60 /200
Game step 40 /200
Game step 50 /200
Game step 30 /200
Game step 70 /200
Game step 40 /200
Game step 40 /200
Game step 60 /200
Game step 70 /200
Game step 60 /200
Game step 40 /200
Game step 50 /200
Game step 80 /200
Game step 50 /200
Game step 70 /200
Game step 50 /200
Game step 70 /200
Game step 80 /200
Game step 50 /200
Game step 60 /200
Game step 90 /200
Game step 60 /200
Game step 80 /200
Game step 60 /200
Game step 90 /200
Game step 80 /200
Game step 60 /200
Game step 70 /200
Game step 90 /200
Game step 70 /200
Game step 100 /200
Game step 70 /200
Game step 100 /200
Terminated due to peaceful activity
Terminated at step 101
Second player won
7[[ 2]
6 [ 2 ]
5 [ 2 2 ]
4 [ 2 -1 ]
3 [ -1 -2 ]
2 [ -2 -2 ]
1 [ -1 -2 -1 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Game step 70 /200
Game step 90 /200
Terminated due to peaceful activity
Terminated at step 90
Second player won
7[[ -1]
6 [ 1 ]
5 [ 2 ]
4 [-1 2 -2 ]
3 [ 1 -2 ]
2 [ 1 -2 ]
1 [ 1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Terminated due to peaceful activity
Terminated at step 106
First player won.
7[[ 2]
6 [ 2 1 ]
5 [ 2 ]
4 [ 2 ]
3 [ -1 -2 ]
2 [-1 -2 ]
1 [ -1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Game step 80 /200
Game step 100 /200
Game step 80 /200
Game step 80 /200
Terminated due to peaceful activity
Terminated at step 104
Second player won
7[[ -1]
6 [ 2 1 ]
5 [ 1 2 ]
4 [ 2 1 ]
3 [ ]
2 [ -2 -1 -2 ]
1 [ -1 -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Terminated due to peaceful activity
Terminated at step 77
First player won.
7[[ 2]
6 [ 2 1 ]
5 [ 2 ]
4 [ 2 ]
3 [ -1 ]
2 [-1 -2 -1 ]
1 [ -1 -2 -1]
0 [ 1 1 ]]
0 1 2 3 4 5 6 7
Game step 90 /200
Game step 90 /200
Game step 90 /200
Terminated due to peaceful activity
Terminated at step 92
Second player won
7[[ -1]
6 [ 2 -1 ]
5 [ 2 2 ]
4 [-1 -1 ]
3 [ 1 -2 1]
2 [ -2 ]
1 [ 1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Game step 100 /200
Game step 100 /200
Terminated due to peaceful activity
Terminated at step 101
Second player won
7[[ -1]
6 [ 2 -1 ]
5 [ 2 -1]
4 [ 2 -2 -1 ]
3 [ 1 -2 1]
2 [ 1 1 ]
1 [ ]
0 [ -2 ]]
0 1 2 3 4 5 6 7
Game step 110 /200
Game step 120 /200
Terminated due to peaceful activity
Terminated at step 127
Second player won
7[[ 2]
6 [ 2 2 1 ]
5 [ 1 2 ]
4 [ ]
3 [ -1 -1 -2 ]
2 [ -2 -1 ]
1 [ 1 -2 -1 ]
0 [ 1 -2 ]]
0 1 2 3 4 5 6 7
MCTS pool has joined
Terminal sentinel is put on queue
Sentinel received. GPU will process this batch and terminate afterwards
Queue task done signal sent. Queue will join. Thread may still be running.
Queue has joined
GPU Thread has joined
Successful generation of many games?
Queue empty: True
lowpuct train epoch 7, batch 0. running value loss: 0.03906. running policy loss: 0.06010. running p diff: 0.04204
lowpuct valid epoch 7, batch 0. validation value loss: 0.03176. validation policy loss: 0.06093 validation p diff: 0.04044
saved model lowpuct at D:\Git\alphazero-checker\saves\lowpuct_7_0.pkl
lowpuct train epoch 7, batch 10. running value loss: 0.02974. running policy loss: 0.05923. running p diff: 0.04290
lowpuct train epoch 7, batch 20. running value loss: 0.03037. running policy loss: 0.05998. running p diff: 0.04272
lowpuct train epoch 7, batch 30. running value loss: 0.03027. running policy loss: 0.06042. running p diff: 0.04256
lowpuct train epoch 7, batch 40. running value loss: 0.03030. running policy loss: 0.06036. running p diff: 0.04192
lowpuct train epoch 7, batch 50. running value loss: 0.03012. running policy loss: 0.06037. running p diff: 0.04180
lowpuct train epoch 7, batch 60. running value loss: 0.03066. running policy loss: 0.06040. running p diff: 0.04147
lowpuct train epoch 7, batch 70. running value loss: 0.03010. running policy loss: 0.06034. running p diff: 0.04143
lowpuct train epoch 7, batch 80. running value loss: 0.02988. running policy loss: 0.06010. running p diff: 0.04128
lowpuct train epoch 7, batch 90. running value loss: 0.03039. running policy loss: 0.06014. running p diff: 0.04120
lowpuct train epoch 7, batch 100. running value loss: 0.02994. running policy loss: 0.05998. running p diff: 0.04111
lowpuct train epoch 7, batch 110. running value loss: 0.02988. running policy loss: 0.06031. running p diff: 0.04132
lowpuct train epoch 7, batch 120. running value loss: 0.03045. running policy loss: 0.06012. running p diff: 0.04092
lowpuct train epoch 7, batch 130. running value loss: 0.03146. running policy loss: 0.06053. running p diff: 0.04082
lowpuct train epoch 7, batch 140. running value loss: 0.03090. running policy loss: 0.06075. running p diff: 0.04120
lowpuct train epoch 7, batch 150. running value loss: 0.03143. running policy loss: 0.06061. running p diff: 0.04108
lowpuct train epoch 7, batch 160. running value loss: 0.03107. running policy loss: 0.06054. running p diff: 0.04101
lowpuct train epoch 7, batch 170. running value loss: 0.03092. running policy loss: 0.06066. running p diff: 0.04141
lowpuct train epoch 7, batch 180. running value loss: 0.03055. running policy loss: 0.06017. running p diff: 0.04150
lowpuct train epoch 7, batch 190. running value loss: 0.03092. running policy loss: 0.06036. running p diff: 0.04145
lowpuct train epoch 7, batch 200. running value loss: 0.03092. running policy loss: 0.06056. running p diff: 0.04133
lowpuct train epoch 7, batch 210. running value loss: 0.03186. running policy loss: 0.06059. running p diff: 0.04140
lowpuct train epoch 7, batch 220. running value loss: 0.03191. running policy loss: 0.06065. running p diff: 0.04086
lowpuct train epoch 7, batch 230. running value loss: 0.03160. running policy loss: 0.06075. running p diff: 0.04044
lowpuct train epoch 7, batch 240. running value loss: 0.03139. running policy loss: 0.06015. running p diff: 0.04030
lowpuct train epoch 7, batch 250. running value loss: 0.03162. running policy loss: 0.06017. running p diff: 0.04031
lowpuct train epoch 7, batch 260. running value loss: 0.03145. running policy loss: 0.06021. running p diff: 0.03986
lowpuct train epoch 7, batch 270. running value loss: 0.03176. running policy loss: 0.06026. running p diff: 0.04050
lowpuct train epoch 7, batch 280. running value loss: 0.03170. running policy loss: 0.06023. running p diff: 0.04100
lowpuct train epoch 7, batch 290. running value loss: 0.03146. running policy loss: 0.06043. running p diff: 0.04095
lowpuct train epoch 7, batch 300. running value loss: 0.03178. running policy loss: 0.06094. running p diff: 0.04126
lowpuct train epoch 7, batch 310. running value loss: 0.03170. running policy loss: 0.06100. running p diff: 0.04175
lowpuct train epoch 7, batch 320. running value loss: 0.03209. running policy loss: 0.06080. running p diff: 0.04119
lowpuct train epoch 7, batch 330. running value loss: 0.03252. running policy loss: 0.06076. running p diff: 0.04122
lowpuct train epoch 7, batch 340. running value loss: 0.03275. running policy loss: 0.06070. running p diff: 0.04136
lowpuct train epoch 7, batch 350. running value loss: 0.03260. running policy loss: 0.06039. running p diff: 0.04131
lowpuct train epoch 7, batch 360. running value loss: 0.03209. running policy loss: 0.06035. running p diff: 0.04148
lowpuct train epoch 7, batch 370. running value loss: 0.03150. running policy loss: 0.06032. running p diff: 0.04192
lowpuct train epoch 7, batch 380. running value loss: 0.03094. running policy loss: 0.06067. running p diff: 0.04194
lowpuct train epoch 7, batch 390. running value loss: 0.03150. running policy loss: 0.06064. running p diff: 0.04206
lowpuct train epoch 7, batch 400. running value loss: 0.03154. running policy loss: 0.06061. running p diff: 0.04196
lowpuct train epoch 7, batch 410. running value loss: 0.03170. running policy loss: 0.06043. running p diff: 0.04145
lowpuct train epoch 7, batch 420. running value loss: 0.03201. running policy loss: 0.06079. running p diff: 0.04112
lowpuct train epoch 7, batch 430. running value loss: 0.03266. running policy loss: 0.06047. running p diff: 0.04128
lowpuct train epoch 7, batch 440. running value loss: 0.03223. running policy loss: 0.06055. running p diff: 0.04136
lowpuct train epoch 7, batch 450. running value loss: 0.03184. running policy loss: 0.06014. running p diff: 0.04125
lowpuct train epoch 7, batch 460. running value loss: 0.03203. running policy loss: 0.06023. running p diff: 0.04127
lowpuct train epoch 7, batch 470. running value loss: 0.03131. running policy loss: 0.05975. running p diff: 0.04114
lowpuct train epoch 7, batch 480. running value loss: 0.03082. running policy loss: 0.06005. running p diff: 0.04072
lowpuct train epoch 7, batch 490. running value loss: 0.03041. running policy loss: 0.06010. running p diff: 0.04059
Generating a new game with MCTS
Generating a new game with MCTS
Generating a new game with MCTS
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Generating a new game with MCTS
Game step 0 /200
Game step 0 /200
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Game step 0 /200
Game step 0 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 20 /200
Terminated at step 22
Second player won
7[[ -1 ]
6 [ -1 -1 -1 ]
5 [ -1 -1 -1 -1]
4 [-1 -1 -1 -1 ]
3 [ 1 1 1 1]
2 [ 1 1 1 1 ]
1 [ 1 1 1]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 10 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 20 /200
Game step 30 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 30 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 40 /200
Game step 50 /200
Game step 50 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 50 /200
Game step 70 /200
Game step 60 /200
Game step 70 /200
Game step 60 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 60 /200
Game step 80 /200
Game step 80 /200
Game step 70 /200
Game step 70 /200
Game step 80 /200
Game step 80 /200
Game step 80 /200
Terminated due to peaceful activity
Terminated at step 87
Second player won
7[[ -1]
6 [ 2 1 ]
5 [ 2 ]
4 [ 2 -2 ]
3 [ -1 ]
2 [-1 -2 ]
1 [ -1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 70 /200
Game step 90 /200
Game step 80 /200
Terminated due to peaceful activity
Terminated at step 87
First player won.
7[[ 2]
6 [ 2 1 ]
5 [ 1 2 ]
4 [ 2 -2 ]
3 [ ]
2 [-1 -2 ]
1 [ -1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 80 /200
Game step 90 /200
Game step 10 /200
Game step 90 /200
Game step 100 /200
Terminated due to peaceful activity
Terminated at step 84
First player won.
7[[ -1]
6 [ 2 2 1 ]
5 [ -1 1 2 1]
4 [ 1 ]
3 [ 1 -2 ]
2 [ -2 -2 ]
1 [ ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 80 /200
Game step 10 /200
Terminated due to peaceful activity
Terminated at step 87
Second player won
7[[ -1 -1]
6 [ 2 2 -1 ]
5 [ 2 1]
4 [ 2 1 ]
3 [ -1 -2 ]
2 [-1 -2 ]
1 [ -1 -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 100 /200
Game step 10 /200
Terminated due to peaceful activity
Terminated at step 109
Second player won
7[[ -1 ]
6 [ 2 -1 ]
5 [ 1 2 ]
4 [ 2 1 ]
3 [ 1]
2 [ -2 -1 ]
1 [ -1 -2 -2]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 10 /200
Game step 90 /200
Game step 100 /200
Game step 30 /200
Game step 10 /200
Game step 110 /200
Game step 30 /200
Terminated due to peaceful activity
Terminated at step 110
Second player won
7[[ 2 2]
6 [ 1 ]
5 [ 2 ]
4 [ -2 ]
3 [ ]
2 [-1 -2 ]
1 [ -1 -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 20 /200
Game step 40 /200
Game step 100 /200
Game step 110 /200
Game step 20 /200
Game step 40 /200
Terminated due to peaceful activity
Terminated at step 110
First player won.
7[[ 2 2]
6 [ 1 ]
5 [ 2 1]
4 [ ]
3 [ -1 2 ]
2 [ -2 -2 ]
1 [ 1 -2 -1 ]
0 [ 1 -2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 10 /200
Game step 30 /200
Game step 30 /200
Terminated due to peaceful activity
Terminated at step 106
Second player won
7[[ 2]
6 [ 2 1 ]
5 [ 2 1 2 ]
4 [ ]
3 [ -1 -2 ]
2 [ -2 -2 ]
1 [ -1 -2 -1 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Game step 50 /200
Game step 30 /200
Game step 10 /200
Game step 20 /200
Game step 50 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 60 /200
Game step 20 /200
Game step 30 /200
Game step 60 /200
Game step 50 /200
Game step 50 /200
Game step 30 /200
Game step 50 /200
Game step 70 /200
Game step 40 /200
Game step 60 /200
Game step 70 /200
Game step 40 /200
Game step 60 /200
Game step 60 /200
Game step 80 /200
Game step 70 /200
Game step 50 /200
Game step 50 /200
Game step 80 /200
Game step 70 /200
Game step 70 /200
Terminated due to peaceful activity
Terminated at step 85
Second player won
7[[ -1 ]
6 [ 1 2 -1 ]
5 [ 2 1]
4 [ 2 1 1 ]
3 [ 1]
2 [ -2 -1 -2 ]
1 [ -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Game step 80 /200
Game step 90 /200
Game step 60 /200
Game step 60 /200
Game step 80 /200
Game step 80 /200
Game step 90 /200
Terminated due to peaceful activity
Terminated at step 98
First player won.
7[[ 2 -1 -1]
6 [ 2 -1 ]
5 [ 2 2 ]
4 [ 1 2 1 1 ]
3 [ -2 ]
2 [ -2 ]
1 [ ]
0 [ -2 ]]
0 1 2 3 4 5 6 7
Game step 90 /200
Game step 70 /200
Game step 70 /200
Game step 90 /200
Terminated due to peaceful activity
Terminated at step 99
First player won.
7[[ -1]
6 [ 2 1 ]
5 [ 2 2 1]
4 [ 1 ]
3 [ 1]
2 [-1 -2 -2 ]
1 [ -1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Game step 100 /200
Game step 80 /200
Terminated due to peaceful activity
Terminated at step 102
Second player won
7[[ -1]
6 [ 1 ]
5 [ 2 ]
4 [ 2 ]
3 [ -1 -1 ]
2 [ -2 -1 -2 ]
1 [ 1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Game step 80 /200
Game step 100 /200
Terminated due to peaceful activity
Terminated at step 102
First player won.
7[[ -1 -1]
6 [ 2 2 1 ]
5 [ 1 1 2 1]
4 [ 1 1 ]
3 [ 1]
2 [ -2 -1 -2 ]
1 [ -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Terminated due to peaceful activity
Terminated at step 87
First player won.
7[[ 2 -1 -1]
6 [ 2 2 -1 ]
5 [ 1 2 ]
4 [ 1 1 ]
3 [ ]
2 [ -2 ]
1 [ -1 ]
0 [ 1 -2 ]]
0 1 2 3 4 5 6 7
Game step 90 /200
Game step 100 /200
Terminated due to peaceful activity
Terminated at step 101
First player won.
7[[ 2]
6 [ 2 ]
5 [ 1 2 ]
4 [-1 ]
3 [ -1 2 -2 ]
2 [ -2 ]
1 [ 1 -2 -2]
0 [ 1 1 ]]
0 1 2 3 4 5 6 7
MCTS pool has joined
Terminal sentinel is put on queue
Sentinel received. GPU will process this batch and terminate afterwards
Queue task done signal sent. Queue will join. Thread may still be running.
Queue has joined
GPU Thread has joined
Successful generation of many games?
Queue empty: True
lowpuct train epoch 8, batch 0. running value loss: 0.03070. running policy loss: 0.06028. running p diff: 0.04063
lowpuct valid epoch 8, batch 0. validation value loss: 0.03211. validation policy loss: 0.05617 validation p diff: 0.03983
saved model lowpuct at D:\Git\alphazero-checker\saves\lowpuct_8_0.pkl
lowpuct train epoch 8, batch 10. running value loss: 0.03025. running policy loss: 0.05949. running p diff: 0.04030
lowpuct train epoch 8, batch 20. running value loss: 0.03105. running policy loss: 0.05898. running p diff: 0.04027
lowpuct train epoch 8, batch 30. running value loss: 0.03084. running policy loss: 0.05824. running p diff: 0.03988
lowpuct train epoch 8, batch 40. running value loss: 0.03131. running policy loss: 0.05770. running p diff: 0.03959
lowpuct train epoch 8, batch 50. running value loss: 0.03102. running policy loss: 0.05707. running p diff: 0.03967
lowpuct train epoch 8, batch 60. running value loss: 0.03082. running policy loss: 0.05755. running p diff: 0.04007
lowpuct train epoch 8, batch 70. running value loss: 0.03131. running policy loss: 0.05749. running p diff: 0.04020
lowpuct train epoch 8, batch 80. running value loss: 0.03123. running policy loss: 0.05706. running p diff: 0.04041
lowpuct train epoch 8, batch 90. running value loss: 0.03150. running policy loss: 0.05700. running p diff: 0.04022
lowpuct train epoch 8, batch 100. running value loss: 0.03193. running policy loss: 0.05721. running p diff: 0.04013
lowpuct train epoch 8, batch 110. running value loss: 0.03266. running policy loss: 0.05668. running p diff: 0.04007
lowpuct train epoch 8, batch 120. running value loss: 0.03223. running policy loss: 0.05686. running p diff: 0.04014
lowpuct train epoch 8, batch 130. running value loss: 0.03252. running policy loss: 0.05704. running p diff: 0.04029
lowpuct train epoch 8, batch 140. running value loss: 0.03221. running policy loss: 0.05684. running p diff: 0.04057
lowpuct train epoch 8, batch 150. running value loss: 0.03145. running policy loss: 0.05658. running p diff: 0.04069
lowpuct train epoch 8, batch 160. running value loss: 0.03117. running policy loss: 0.05636. running p diff: 0.04064
lowpuct train epoch 8, batch 170. running value loss: 0.03080. running policy loss: 0.05629. running p diff: 0.04068
lowpuct train epoch 8, batch 180. running value loss: 0.03061. running policy loss: 0.05635. running p diff: 0.04058
lowpuct train epoch 8, batch 190. running value loss: 0.03143. running policy loss: 0.05663. running p diff: 0.04077
lowpuct train epoch 8, batch 200. running value loss: 0.03229. running policy loss: 0.05700. running p diff: 0.04079
lowpuct train epoch 8, batch 210. running value loss: 0.03232. running policy loss: 0.05716. running p diff: 0.04060
lowpuct train epoch 8, batch 220. running value loss: 0.03205. running policy loss: 0.05716. running p diff: 0.04034
lowpuct train epoch 8, batch 230. running value loss: 0.03252. running policy loss: 0.05686. running p diff: 0.04057
lowpuct train epoch 8, batch 240. running value loss: 0.03146. running policy loss: 0.05667. running p diff: 0.04044
lowpuct train epoch 8, batch 250. running value loss: 0.03090. running policy loss: 0.05624. running p diff: 0.04068
lowpuct train epoch 8, batch 260. running value loss: 0.03043. running policy loss: 0.05648. running p diff: 0.04080
lowpuct train epoch 8, batch 270. running value loss: 0.03049. running policy loss: 0.05636. running p diff: 0.04088
lowpuct train epoch 8, batch 280. running value loss: 0.03064. running policy loss: 0.05651. running p diff: 0.04047
lowpuct train epoch 8, batch 290. running value loss: 0.03062. running policy loss: 0.05637. running p diff: 0.04074
lowpuct train epoch 8, batch 300. running value loss: 0.03123. running policy loss: 0.05692. running p diff: 0.04075
lowpuct train epoch 8, batch 310. running value loss: 0.03164. running policy loss: 0.05698. running p diff: 0.04081
lowpuct train epoch 8, batch 320. running value loss: 0.03146. running policy loss: 0.05735. running p diff: 0.04095
lowpuct train epoch 8, batch 330. running value loss: 0.03100. running policy loss: 0.05796. running p diff: 0.04108
lowpuct train epoch 8, batch 340. running value loss: 0.03094. running policy loss: 0.05782. running p diff: 0.04098
lowpuct train epoch 8, batch 350. running value loss: 0.03045. running policy loss: 0.05712. running p diff: 0.04071
lowpuct train epoch 8, batch 360. running value loss: 0.03121. running policy loss: 0.05669. running p diff: 0.04065
lowpuct train epoch 8, batch 370. running value loss: 0.03172. running policy loss: 0.05649. running p diff: 0.04055
lowpuct train epoch 8, batch 380. running value loss: 0.03164. running policy loss: 0.05598. running p diff: 0.04059
lowpuct train epoch 8, batch 390. running value loss: 0.03213. running policy loss: 0.05617. running p diff: 0.04053
lowpuct train epoch 8, batch 400. running value loss: 0.03238. running policy loss: 0.05616. running p diff: 0.04043
lowpuct train epoch 8, batch 410. running value loss: 0.03139. running policy loss: 0.05604. running p diff: 0.04021
lowpuct train epoch 8, batch 420. running value loss: 0.03098. running policy loss: 0.05613. running p diff: 0.03994
lowpuct train epoch 8, batch 430. running value loss: 0.03188. running policy loss: 0.05595. running p diff: 0.04017
lowpuct train epoch 8, batch 440. running value loss: 0.03168. running policy loss: 0.05611. running p diff: 0.03997
lowpuct train epoch 8, batch 450. running value loss: 0.03191. running policy loss: 0.05625. running p diff: 0.03954
lowpuct train epoch 8, batch 460. running value loss: 0.03229. running policy loss: 0.05691. running p diff: 0.03991
lowpuct train epoch 8, batch 470. running value loss: 0.03232. running policy loss: 0.05688. running p diff: 0.04016
lowpuct train epoch 8, batch 480. running value loss: 0.03133. running policy loss: 0.05699. running p diff: 0.04041
lowpuct train epoch 8, batch 490. running value loss: 0.03094. running policy loss: 0.05675. running p diff: 0.04059
Generating a new game with MCTS
Generating a new game with MCTS
Generating a new game with MCTS
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Generating a new game with MCTS
Game step 0 /200
Game step 0 /200
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Game step 0 /200
Game step 0 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 30 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Terminated due to peaceful activity
Terminated at step 77
Second player won
7[[ 2]
6 [ 1 2 1 ]
5 [ 2 1]
4 [ ]
3 [ ]
2 [ -2 -1 ]
1 [ -1 -2 ]
0 [-2 -2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 70 /200
Game step 80 /200
Terminated due to peaceful activity
Terminated at step 75
Second player won
7[[ 2 -1]
6 [ 2 1 ]
5 [ 2 1]
4 [ 1 ]
3 [ -1 ]
2 [-1 -2 ]
1 [ -1 -2 -2]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 80 /200
Game step 80 /200
Game step 80 /200
Game step 10 /200
Game step 80 /200
Game step 90 /200
Game step 10 /200
Game step 80 /200
Game step 90 /200
Terminated due to peaceful activity
Terminated at step 95
First player won.
7[[ 2]
6 [ 2 2 1 ]
5 [ 1 2 ]
4 [ 1 ]
3 [ ]
2 [-1 -2 -2 ]
1 [ -1 -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 20 /200
Game step 90 /200
Game step 90 /200
Game step 90 /200
Game step 10 /200
Terminated due to peaceful activity
Terminated at step 94
First player won.
7[[ 2]
6 [ 2 ]
5 [ 2 2 ]
4 [ -1 ]
3 [ -1 1 1 ]
2 [ -2 -2 ]
1 [ 1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 30 /200
Game step 90 /200
Game step 100 /200
Game step 30 /200
Terminated due to peaceful activity
Terminated at step 103
Second player won
7[[ -1]
6 [ 2 1 ]
5 [ 2 2 1]
4 [ 1 ]
3 [ -1 ]
2 [-1 -2 -2 ]
1 [ -1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 100 /200
Game step 100 /200
Game step 10 /200
Game step 40 /200
Game step 40 /200
Game step 50 /200
Terminated at step 51
First player won.
7[[ 2]
6 [ 1 ]
5 [ 1 2 ]
4 [ 1 ]
3 [ ]
2 [-1 2 1 ]
1 [ 1 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 10 /200
Game step 100 /200
Game step 20 /200
Game step 30 /200
Game step 110 /200
Game step 10 /200
Game step 110 /200
Game step 50 /200
Terminated due to peaceful activity
Terminated at step 111
Second player won
7[[ 2 -1]
6 [ 1 ]
5 [ 2 ]
4 [ 1 -2 ]
3 [ 2 -2 ]
2 [-1 -2 ]
1 [ 1 -2 -1 ]
0 [ 1 -2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 30 /200
Terminated due to peaceful activity
Terminated at step 108
First player won.
7[[ -1]
6 [ 2 2 1 ]
5 [ 2 1 2 ]
4 [ 1 1 ]
3 [ 1 -2 ]
2 [ -2 -2 ]
1 [ -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 120 /200
Game step 40 /200
Game step 10 /200
Game step 30 /200
Game step 60 /200
Game step 10 /200
Game step 40 /200
Game step 40 /200
Game step 130 /200
Game step 30 /200
Game step 20 /200
Game step 20 /200
Terminated at step 22
Second player won
7[[ -1]
6 [-1 -1 -1 ]
5 [ -1 -1 -1 -1]
4 [-1 -1 -1 -1 ]
3 [ 1 1 1 1]
2 [ 1 1 1 1 ]
1 [ 1 1 1 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Game step 50 /200
Game step 70 /200
Terminated due to peaceful activity
Terminated at step 134
First player won.
7[[ 2]
6 [ 2 1 ]
5 [ 2 2 ]
4 [ -2 ]
3 [ 2 -2 ]
2 [ -2 -2 ]
1 [ -1 -2 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Game step 30 /200
Game step 50 /200
Game step 50 /200
Game step 40 /200
Game step 60 /200
Game step 40 /200
Game step 80 /200
Game step 60 /200
Game step 60 /200
Game step 50 /200
Game step 50 /200
Game step 70 /200
Game step 90 /200
Game step 70 /200
Terminated due to peaceful activity
Terminated at step 90
First player won.
7[[ -1]
6 [ 1 ]
5 [ 1 2 1]
4 [ 2 1 ]
3 [ ]
2 [-1 -2 -1 ]
1 [ -1 -2 ]
0 [ 1 1 ]]
0 1 2 3 4 5 6 7
Game step 70 /200
Game step 60 /200
Game step 60 /200
Game step 80 /200
Game step 80 /200
Game step 80 /200
Game step 70 /200
Game step 70 /200
Game step 90 /200
Game step 90 /200
Terminated due to peaceful activity
Terminated at step 95
Second player won
7[[ -1]
6 [ 2 1 ]
5 [ 2 ]
4 [ ]
3 [ -2 ]
2 [-1 -2 -1 ]
1 [ -1 -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Game step 90 /200
Game step 80 /200
Terminated due to peaceful activity
Terminated at step 90
First player won.
7[[ 2 -1 -1]
6 [ 2 2 -1 ]
5 [ 2 -2 2 ]
4 [ 1 ]
3 [ -1 1]
2 [-1 -2 ]
1 [ -1 ]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Game step 80 /200
Terminated due to peaceful activity
Terminated at step 89
Second player won
7[[ 2]
6 [ 2 1 ]
5 [ 2 1]
4 [ 2 ]
3 [ -1 -2 ]
2 [-1 -2 ]
1 [ -1 -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Game step 100 /200
Terminated due to peaceful activity
Terminated at step 104
Second player won
7[[ 2 -1 ]
6 [ 1 2 -1 ]
5 [ 2 1]
4 [ 2 -2 ]
3 [ ]
2 [ -2 -2 ]
1 [ -2 ]
0 [-2 ]]
0 1 2 3 4 5 6 7
Game step 90 /200
Game step 100 /200
Terminated due to peaceful activity
Terminated at step 104
First player won.
7[[ 2 -1 -1]
6 [ 2 2 -1 ]
5 [ 2 1 2 ]
4 [ 2 -2 1 ]
3 [ ]
2 [ -2 ]
1 [ -2 ]
0 [ ]]
0 1 2 3 4 5 6 7
MCTS pool has joined
Terminal sentinel is put on queue
Sentinel received. GPU will process this batch and terminate afterwards
Queue task done signal sent. Queue will join. Thread may still be running.
Queue has joined
GPU Thread has joined
Successful generation of many games?
Queue empty: True
lowpuct train epoch 9, batch 0. running value loss: 0.03084. running policy loss: 0.05677. running p diff: 0.04116
lowpuct valid epoch 9, batch 0. validation value loss: 0.03175. validation policy loss: 0.05418 validation p diff: 0.03996
saved model lowpuct at D:\Git\alphazero-checker\saves\lowpuct_9_0.pkl
lowpuct train epoch 9, batch 10. running value loss: 0.03080. running policy loss: 0.05640. running p diff: 0.04120
lowpuct train epoch 9, batch 20. running value loss: 0.03119. running policy loss: 0.05609. running p diff: 0.04132
lowpuct train epoch 9, batch 30. running value loss: 0.03193. running policy loss: 0.05580. running p diff: 0.04126
lowpuct train epoch 9, batch 40. running value loss: 0.03244. running policy loss: 0.05567. running p diff: 0.04086
lowpuct train epoch 9, batch 50. running value loss: 0.03215. running policy loss: 0.05575. running p diff: 0.04033
lowpuct train epoch 9, batch 60. running value loss: 0.03176. running policy loss: 0.05571. running p diff: 0.03987
lowpuct train epoch 9, batch 70. running value loss: 0.03150. running policy loss: 0.05556. running p diff: 0.03954
lowpuct train epoch 9, batch 80. running value loss: 0.03111. running policy loss: 0.05542. running p diff: 0.03918
lowpuct train epoch 9, batch 90. running value loss: 0.03119. running policy loss: 0.05507. running p diff: 0.03965
lowpuct train epoch 9, batch 100. running value loss: 0.03133. running policy loss: 0.05482. running p diff: 0.04020
lowpuct train epoch 9, batch 110. running value loss: 0.03180. running policy loss: 0.05476. running p diff: 0.04023
lowpuct train epoch 9, batch 120. running value loss: 0.03182. running policy loss: 0.05471. running p diff: 0.04054
lowpuct train epoch 9, batch 130. running value loss: 0.03156. running policy loss: 0.05446. running p diff: 0.04053
lowpuct train epoch 9, batch 140. running value loss: 0.03146. running policy loss: 0.05432. running p diff: 0.04013
lowpuct train epoch 9, batch 150. running value loss: 0.03150. running policy loss: 0.05419. running p diff: 0.04011
lowpuct train epoch 9, batch 160. running value loss: 0.03090. running policy loss: 0.05421. running p diff: 0.04018
lowpuct train epoch 9, batch 170. running value loss: 0.03098. running policy loss: 0.05410. running p diff: 0.04007
lowpuct train epoch 9, batch 180. running value loss: 0.03166. running policy loss: 0.05440. running p diff: 0.04003
lowpuct train epoch 9, batch 190. running value loss: 0.03123. running policy loss: 0.05461. running p diff: 0.04033
lowpuct train epoch 9, batch 200. running value loss: 0.03092. running policy loss: 0.05499. running p diff: 0.04040
lowpuct train epoch 9, batch 210. running value loss: 0.03104. running policy loss: 0.05443. running p diff: 0.04068
lowpuct train epoch 9, batch 220. running value loss: 0.03082. running policy loss: 0.05473. running p diff: 0.04082
lowpuct train epoch 9, batch 230. running value loss: 0.02996. running policy loss: 0.05470. running p diff: 0.04034
lowpuct train epoch 9, batch 240. running value loss: 0.03057. running policy loss: 0.05484. running p diff: 0.04023
lowpuct train epoch 9, batch 250. running value loss: 0.03045. running policy loss: 0.05445. running p diff: 0.04023
lowpuct train epoch 9, batch 260. running value loss: 0.03082. running policy loss: 0.05473. running p diff: 0.04023
lowpuct train epoch 9, batch 270. running value loss: 0.03139. running policy loss: 0.05417. running p diff: 0.04029
lowpuct train epoch 9, batch 280. running value loss: 0.03170. running policy loss: 0.05405. running p diff: 0.04083
lowpuct train epoch 9, batch 290. running value loss: 0.03135. running policy loss: 0.05410. running p diff: 0.04094
lowpuct train epoch 9, batch 300. running value loss: 0.03119. running policy loss: 0.05402. running p diff: 0.04093
lowpuct train epoch 9, batch 310. running value loss: 0.03137. running policy loss: 0.05408. running p diff: 0.04110
lowpuct train epoch 9, batch 320. running value loss: 0.03092. running policy loss: 0.05444. running p diff: 0.04087
lowpuct train epoch 9, batch 330. running value loss: 0.03055. running policy loss: 0.05451. running p diff: 0.04065
lowpuct train epoch 9, batch 340. running value loss: 0.03092. running policy loss: 0.05392. running p diff: 0.04064
lowpuct train epoch 9, batch 350. running value loss: 0.03139. running policy loss: 0.05401. running p diff: 0.04036
lowpuct train epoch 9, batch 360. running value loss: 0.03105. running policy loss: 0.05428. running p diff: 0.04032
lowpuct train epoch 9, batch 370. running value loss: 0.03148. running policy loss: 0.05422. running p diff: 0.04042
lowpuct train epoch 9, batch 380. running value loss: 0.03131. running policy loss: 0.05422. running p diff: 0.04068
lowpuct train epoch 9, batch 390. running value loss: 0.03088. running policy loss: 0.05429. running p diff: 0.04071
lowpuct train epoch 9, batch 400. running value loss: 0.03053. running policy loss: 0.05411. running p diff: 0.04090
lowpuct train epoch 9, batch 410. running value loss: 0.03043. running policy loss: 0.05367. running p diff: 0.04088
lowpuct train epoch 9, batch 420. running value loss: 0.02994. running policy loss: 0.05354. running p diff: 0.04090
lowpuct train epoch 9, batch 430. running value loss: 0.03066. running policy loss: 0.05340. running p diff: 0.04097
lowpuct train epoch 9, batch 440. running value loss: 0.03113. running policy loss: 0.05380. running p diff: 0.04112
lowpuct train epoch 9, batch 450. running value loss: 0.03127. running policy loss: 0.05412. running p diff: 0.04096
lowpuct train epoch 9, batch 460. running value loss: 0.03164. running policy loss: 0.05437. running p diff: 0.04091
lowpuct train epoch 9, batch 470. running value loss: 0.03141. running policy loss: 0.05418. running p diff: 0.04091
lowpuct train epoch 9, batch 480. running value loss: 0.03115. running policy loss: 0.05431. running p diff: 0.04117
lowpuct train epoch 9, batch 490. running value loss: 0.03006. running policy loss: 0.05442. running p diff: 0.04090
Generating a new game with MCTS
Generating a new game with MCTS
Generating a new game with MCTS
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Generating a new game with MCTS
Generating a new game with MCTS
Game step 0 /200
Game step 0 /200
Game step 0 /200
Generating a new game with MCTS
Game step 0 /200
Game step 0 /200
Game step 0 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 10 /200
Game step 20 /200
Terminated at step 22
Second player won
7[[ -1 ]
6 [-1 -1 -1 ]
5 [ -1 -1 -1 -1]
4 [-1 -1 -1 -1 ]
3 [ 1 1 1 1]
2 [ 1 1 1 1 ]
1 [ 1 1 1]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Terminated at step 22
Second player won
7[[ -1]
6 [-1 -1 -1 ]
5 [ -1 -1 -1 -1]
4 [-1 -1 -1 -1 ]
3 [ 1 1 1 1]
2 [ 1 1 1 1 ]
1 [ 1 1 1]
0 [ 1 ]]
0 1 2 3 4 5 6 7
Generating a new game with MCTS
Game step 0 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 20 /200
Game step 30 /200
Game step 20 /200
Game step 10 /200
Game step 10 /200
Game step 30 /200
Game step 30 /200
Game step 40 /200
Game step 30 /200
Game step 20 /200
Game step 30 /200
Game step 20 /200
Game step 30 /200
Game step 30 /200
Game step 40 /200
Game step 40 /200
Game step 50 /200
Game step 30 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 40 /200
Game step 50 /200
Game step 60 /200
Game step 50 /200
Game step 50 /200
Game step 40 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 50 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 70 /200
Game step 60 /200
Game step 60 /200
Game step 60 /200
Game step 70 /200
Game step 60 /200
Game step 70 /200
Game step 70 /200
Game step 80 /200
Game step 70 /200
Game step 70 /200
Game step 70 /200
Game step 80 /200
Game step 80 /200
Game step 70 /200
Game step 80 /200
Game step 80 /200
Game step 90 /200
Game step 80 /200
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment