Created
December 2, 2019 00:13
-
-
Save phimachine/a418ff601ef675023956cb22e89f3f7c to your computer and use it in GitHub Desktop.
Current log
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
C:\Users\JasonHu\Anaconda3\envs\alphazero-checker\python.exe "C:\Users\JasonHu\AppData\Local\JetBrains\PyCharm 2018.3\helpers\pydev\pydevconsole.py" --mode=client --port=13753 | |
import sys; print('Python %s on %s' % (sys.version, sys.platform)) | |
sys.path.extend(['D:\\Git\\alphazero-checker', 'D:/Git/alphazero-checker']) | |
PyDev console: starting. | |
Python 3.6.8 |Anaconda, Inc.| (default, Dec 30 2018, 18:50:55) [MSC v.1915 64 bit (AMD64)] on win32 | |
runfile('D:/Git/alphazero-checker/zero.py', wdir='D:/Git/alphazero-checker') | |
loading model at D:\Git\alphazero-checker\saves\lowpuct_7_0.pkl | |
Loaded model at epoch 7 iteration 0 | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 60 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 80 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 90 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Terminated due to peaceful activity | |
Terminated at step 82 | |
First player won. | |
7[[ 2 -1 -1] | |
6 [ 2 -1 ] | |
5 [ 2 1 2 ] | |
4 [ 1 ] | |
3 [ 1] | |
2 [ -2 -2 ] | |
1 [ -1 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Terminated due to peaceful activity | |
Terminated at step 89 | |
Second player won | |
7[[ -1] | |
6 [ 2 1 ] | |
5 [ 2 2 1] | |
4 [ 1 ] | |
3 [ -1 ] | |
2 [-1 -2 -2 ] | |
1 [ -1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 80 /200 | |
Game step 100 /200 | |
Game step 90 /200 | |
Game step 90 /200 | |
Game step 90 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 90 /200 | |
Game step 90 /200 | |
Terminated due to peaceful activity | |
Terminated at step 94 | |
First player won. | |
7[[ -1 -1] | |
6 [ 2 2 1 ] | |
5 [ 1 2 1] | |
4 [ -2 ] | |
3 [ ] | |
2 [ -2 ] | |
1 [ -1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 110 /200 | |
Terminated due to peaceful activity | |
Terminated at step 111 | |
Second player won | |
7[[ 2] | |
6 [ 2 1 ] | |
5 [ 1 2 ] | |
4 [ 2 ] | |
3 [ -2 ] | |
2 [-1 -2 ] | |
1 [ -1 -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 100 /200 | |
Game step 20 /200 | |
Game step 100 /200 | |
Game step 100 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 30 /200 | |
Terminated due to peaceful activity | |
Terminated at step 107 | |
Second player won | |
7[[ -1] | |
6 [ 2 1 ] | |
5 [ 2 1 2 ] | |
4 [ 1 ] | |
3 [ -2 ] | |
2 [-1 -2 ] | |
1 [ -1 -2 -2] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 100 /200 | |
Game step 30 /200 | |
Terminated due to peaceful activity | |
Terminated at step 104 | |
First player won. | |
7[[ 2] | |
6 [ 2 ] | |
5 [ 1 2 ] | |
4 [ 2 1 ] | |
3 [ -1 ] | |
2 [ -2 -1 ] | |
1 [ 1 -2 -2] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 110 /200 | |
Terminated due to peaceful activity | |
Terminated at step 112 | |
First player won. | |
7[[ 2] | |
6 [ 2 1 ] | |
5 [ 2 2 1] | |
4 [-1 ] | |
3 [ -1 -1 -2 ] | |
2 [ -2 -1 ] | |
1 [ 1 -2 ] | |
0 [ 1 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 40 /200 | |
Game step 10 /200 | |
Game step 20 /200 | |
Game step 10 /200 | |
Terminated due to peaceful activity | |
Terminated at step 107 | |
First player won. | |
7[[ 2 -1] | |
6 [ 2 -1 ] | |
5 [ 2 2 -1] | |
4 [ 1 -2 -1 ] | |
3 [ 2 -2 1] | |
2 [ -2 ] | |
1 [ ] | |
0 [ ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 40 /200 | |
Game step 10 /200 | |
Game step 30 /200 | |
Game step 50 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 20 /200 | |
Game step 10 /200 | |
Game step 20 /200 | |
Game step 50 /200 | |
Game step 40 /200 | |
Game step 30 /200 | |
Game step 60 /200 | |
Game step 40 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 50 /200 | |
Game step 60 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 30 /200 | |
Game step 70 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 60 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 80 /200 | |
Game step 50 /200 | |
Game step 70 /200 | |
Game step 50 /200 | |
Game step 70 /200 | |
Game step 80 /200 | |
Game step 50 /200 | |
Game step 60 /200 | |
Game step 90 /200 | |
Game step 60 /200 | |
Game step 80 /200 | |
Game step 60 /200 | |
Game step 90 /200 | |
Game step 80 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 90 /200 | |
Game step 70 /200 | |
Game step 100 /200 | |
Game step 70 /200 | |
Game step 100 /200 | |
Terminated due to peaceful activity | |
Terminated at step 101 | |
Second player won | |
7[[ 2] | |
6 [ 2 ] | |
5 [ 2 2 ] | |
4 [ 2 -1 ] | |
3 [ -1 -2 ] | |
2 [ -2 -2 ] | |
1 [ -1 -2 -1 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 70 /200 | |
Game step 90 /200 | |
Terminated due to peaceful activity | |
Terminated at step 90 | |
Second player won | |
7[[ -1] | |
6 [ 1 ] | |
5 [ 2 ] | |
4 [-1 2 -2 ] | |
3 [ 1 -2 ] | |
2 [ 1 -2 ] | |
1 [ 1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Terminated due to peaceful activity | |
Terminated at step 106 | |
First player won. | |
7[[ 2] | |
6 [ 2 1 ] | |
5 [ 2 ] | |
4 [ 2 ] | |
3 [ -1 -2 ] | |
2 [-1 -2 ] | |
1 [ -1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 80 /200 | |
Game step 100 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Terminated due to peaceful activity | |
Terminated at step 104 | |
Second player won | |
7[[ -1] | |
6 [ 2 1 ] | |
5 [ 1 2 ] | |
4 [ 2 1 ] | |
3 [ ] | |
2 [ -2 -1 -2 ] | |
1 [ -1 -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Terminated due to peaceful activity | |
Terminated at step 77 | |
First player won. | |
7[[ 2] | |
6 [ 2 1 ] | |
5 [ 2 ] | |
4 [ 2 ] | |
3 [ -1 ] | |
2 [-1 -2 -1 ] | |
1 [ -1 -2 -1] | |
0 [ 1 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 90 /200 | |
Game step 90 /200 | |
Game step 90 /200 | |
Terminated due to peaceful activity | |
Terminated at step 92 | |
Second player won | |
7[[ -1] | |
6 [ 2 -1 ] | |
5 [ 2 2 ] | |
4 [-1 -1 ] | |
3 [ 1 -2 1] | |
2 [ -2 ] | |
1 [ 1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 100 /200 | |
Game step 100 /200 | |
Terminated due to peaceful activity | |
Terminated at step 101 | |
Second player won | |
7[[ -1] | |
6 [ 2 -1 ] | |
5 [ 2 -1] | |
4 [ 2 -2 -1 ] | |
3 [ 1 -2 1] | |
2 [ 1 1 ] | |
1 [ ] | |
0 [ -2 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 110 /200 | |
Game step 120 /200 | |
Terminated due to peaceful activity | |
Terminated at step 127 | |
Second player won | |
7[[ 2] | |
6 [ 2 2 1 ] | |
5 [ 1 2 ] | |
4 [ ] | |
3 [ -1 -1 -2 ] | |
2 [ -2 -1 ] | |
1 [ 1 -2 -1 ] | |
0 [ 1 -2 ]] | |
0 1 2 3 4 5 6 7 | |
MCTS pool has joined | |
Terminal sentinel is put on queue | |
Sentinel received. GPU will process this batch and terminate afterwards | |
Queue task done signal sent. Queue will join. Thread may still be running. | |
Queue has joined | |
GPU Thread has joined | |
Successful generation of many games? | |
Queue empty: True | |
lowpuct train epoch 7, batch 0. running value loss: 0.03906. running policy loss: 0.06010. running p diff: 0.04204 | |
lowpuct valid epoch 7, batch 0. validation value loss: 0.03176. validation policy loss: 0.06093 validation p diff: 0.04044 | |
saved model lowpuct at D:\Git\alphazero-checker\saves\lowpuct_7_0.pkl | |
lowpuct train epoch 7, batch 10. running value loss: 0.02974. running policy loss: 0.05923. running p diff: 0.04290 | |
lowpuct train epoch 7, batch 20. running value loss: 0.03037. running policy loss: 0.05998. running p diff: 0.04272 | |
lowpuct train epoch 7, batch 30. running value loss: 0.03027. running policy loss: 0.06042. running p diff: 0.04256 | |
lowpuct train epoch 7, batch 40. running value loss: 0.03030. running policy loss: 0.06036. running p diff: 0.04192 | |
lowpuct train epoch 7, batch 50. running value loss: 0.03012. running policy loss: 0.06037. running p diff: 0.04180 | |
lowpuct train epoch 7, batch 60. running value loss: 0.03066. running policy loss: 0.06040. running p diff: 0.04147 | |
lowpuct train epoch 7, batch 70. running value loss: 0.03010. running policy loss: 0.06034. running p diff: 0.04143 | |
lowpuct train epoch 7, batch 80. running value loss: 0.02988. running policy loss: 0.06010. running p diff: 0.04128 | |
lowpuct train epoch 7, batch 90. running value loss: 0.03039. running policy loss: 0.06014. running p diff: 0.04120 | |
lowpuct train epoch 7, batch 100. running value loss: 0.02994. running policy loss: 0.05998. running p diff: 0.04111 | |
lowpuct train epoch 7, batch 110. running value loss: 0.02988. running policy loss: 0.06031. running p diff: 0.04132 | |
lowpuct train epoch 7, batch 120. running value loss: 0.03045. running policy loss: 0.06012. running p diff: 0.04092 | |
lowpuct train epoch 7, batch 130. running value loss: 0.03146. running policy loss: 0.06053. running p diff: 0.04082 | |
lowpuct train epoch 7, batch 140. running value loss: 0.03090. running policy loss: 0.06075. running p diff: 0.04120 | |
lowpuct train epoch 7, batch 150. running value loss: 0.03143. running policy loss: 0.06061. running p diff: 0.04108 | |
lowpuct train epoch 7, batch 160. running value loss: 0.03107. running policy loss: 0.06054. running p diff: 0.04101 | |
lowpuct train epoch 7, batch 170. running value loss: 0.03092. running policy loss: 0.06066. running p diff: 0.04141 | |
lowpuct train epoch 7, batch 180. running value loss: 0.03055. running policy loss: 0.06017. running p diff: 0.04150 | |
lowpuct train epoch 7, batch 190. running value loss: 0.03092. running policy loss: 0.06036. running p diff: 0.04145 | |
lowpuct train epoch 7, batch 200. running value loss: 0.03092. running policy loss: 0.06056. running p diff: 0.04133 | |
lowpuct train epoch 7, batch 210. running value loss: 0.03186. running policy loss: 0.06059. running p diff: 0.04140 | |
lowpuct train epoch 7, batch 220. running value loss: 0.03191. running policy loss: 0.06065. running p diff: 0.04086 | |
lowpuct train epoch 7, batch 230. running value loss: 0.03160. running policy loss: 0.06075. running p diff: 0.04044 | |
lowpuct train epoch 7, batch 240. running value loss: 0.03139. running policy loss: 0.06015. running p diff: 0.04030 | |
lowpuct train epoch 7, batch 250. running value loss: 0.03162. running policy loss: 0.06017. running p diff: 0.04031 | |
lowpuct train epoch 7, batch 260. running value loss: 0.03145. running policy loss: 0.06021. running p diff: 0.03986 | |
lowpuct train epoch 7, batch 270. running value loss: 0.03176. running policy loss: 0.06026. running p diff: 0.04050 | |
lowpuct train epoch 7, batch 280. running value loss: 0.03170. running policy loss: 0.06023. running p diff: 0.04100 | |
lowpuct train epoch 7, batch 290. running value loss: 0.03146. running policy loss: 0.06043. running p diff: 0.04095 | |
lowpuct train epoch 7, batch 300. running value loss: 0.03178. running policy loss: 0.06094. running p diff: 0.04126 | |
lowpuct train epoch 7, batch 310. running value loss: 0.03170. running policy loss: 0.06100. running p diff: 0.04175 | |
lowpuct train epoch 7, batch 320. running value loss: 0.03209. running policy loss: 0.06080. running p diff: 0.04119 | |
lowpuct train epoch 7, batch 330. running value loss: 0.03252. running policy loss: 0.06076. running p diff: 0.04122 | |
lowpuct train epoch 7, batch 340. running value loss: 0.03275. running policy loss: 0.06070. running p diff: 0.04136 | |
lowpuct train epoch 7, batch 350. running value loss: 0.03260. running policy loss: 0.06039. running p diff: 0.04131 | |
lowpuct train epoch 7, batch 360. running value loss: 0.03209. running policy loss: 0.06035. running p diff: 0.04148 | |
lowpuct train epoch 7, batch 370. running value loss: 0.03150. running policy loss: 0.06032. running p diff: 0.04192 | |
lowpuct train epoch 7, batch 380. running value loss: 0.03094. running policy loss: 0.06067. running p diff: 0.04194 | |
lowpuct train epoch 7, batch 390. running value loss: 0.03150. running policy loss: 0.06064. running p diff: 0.04206 | |
lowpuct train epoch 7, batch 400. running value loss: 0.03154. running policy loss: 0.06061. running p diff: 0.04196 | |
lowpuct train epoch 7, batch 410. running value loss: 0.03170. running policy loss: 0.06043. running p diff: 0.04145 | |
lowpuct train epoch 7, batch 420. running value loss: 0.03201. running policy loss: 0.06079. running p diff: 0.04112 | |
lowpuct train epoch 7, batch 430. running value loss: 0.03266. running policy loss: 0.06047. running p diff: 0.04128 | |
lowpuct train epoch 7, batch 440. running value loss: 0.03223. running policy loss: 0.06055. running p diff: 0.04136 | |
lowpuct train epoch 7, batch 450. running value loss: 0.03184. running policy loss: 0.06014. running p diff: 0.04125 | |
lowpuct train epoch 7, batch 460. running value loss: 0.03203. running policy loss: 0.06023. running p diff: 0.04127 | |
lowpuct train epoch 7, batch 470. running value loss: 0.03131. running policy loss: 0.05975. running p diff: 0.04114 | |
lowpuct train epoch 7, batch 480. running value loss: 0.03082. running policy loss: 0.06005. running p diff: 0.04072 | |
lowpuct train epoch 7, batch 490. running value loss: 0.03041. running policy loss: 0.06010. running p diff: 0.04059 | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 20 /200 | |
Terminated at step 22 | |
Second player won | |
7[[ -1 ] | |
6 [ -1 -1 -1 ] | |
5 [ -1 -1 -1 -1] | |
4 [-1 -1 -1 -1 ] | |
3 [ 1 1 1 1] | |
2 [ 1 1 1 1 ] | |
1 [ 1 1 1] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 10 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 30 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 50 /200 | |
Game step 70 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 60 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Terminated due to peaceful activity | |
Terminated at step 87 | |
Second player won | |
7[[ -1] | |
6 [ 2 1 ] | |
5 [ 2 ] | |
4 [ 2 -2 ] | |
3 [ -1 ] | |
2 [-1 -2 ] | |
1 [ -1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 70 /200 | |
Game step 90 /200 | |
Game step 80 /200 | |
Terminated due to peaceful activity | |
Terminated at step 87 | |
First player won. | |
7[[ 2] | |
6 [ 2 1 ] | |
5 [ 1 2 ] | |
4 [ 2 -2 ] | |
3 [ ] | |
2 [-1 -2 ] | |
1 [ -1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 80 /200 | |
Game step 90 /200 | |
Game step 10 /200 | |
Game step 90 /200 | |
Game step 100 /200 | |
Terminated due to peaceful activity | |
Terminated at step 84 | |
First player won. | |
7[[ -1] | |
6 [ 2 2 1 ] | |
5 [ -1 1 2 1] | |
4 [ 1 ] | |
3 [ 1 -2 ] | |
2 [ -2 -2 ] | |
1 [ ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 80 /200 | |
Game step 10 /200 | |
Terminated due to peaceful activity | |
Terminated at step 87 | |
Second player won | |
7[[ -1 -1] | |
6 [ 2 2 -1 ] | |
5 [ 2 1] | |
4 [ 2 1 ] | |
3 [ -1 -2 ] | |
2 [-1 -2 ] | |
1 [ -1 -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 100 /200 | |
Game step 10 /200 | |
Terminated due to peaceful activity | |
Terminated at step 109 | |
Second player won | |
7[[ -1 ] | |
6 [ 2 -1 ] | |
5 [ 1 2 ] | |
4 [ 2 1 ] | |
3 [ 1] | |
2 [ -2 -1 ] | |
1 [ -1 -2 -2] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 10 /200 | |
Game step 90 /200 | |
Game step 100 /200 | |
Game step 30 /200 | |
Game step 10 /200 | |
Game step 110 /200 | |
Game step 30 /200 | |
Terminated due to peaceful activity | |
Terminated at step 110 | |
Second player won | |
7[[ 2 2] | |
6 [ 1 ] | |
5 [ 2 ] | |
4 [ -2 ] | |
3 [ ] | |
2 [-1 -2 ] | |
1 [ -1 -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 40 /200 | |
Game step 100 /200 | |
Game step 110 /200 | |
Game step 20 /200 | |
Game step 40 /200 | |
Terminated due to peaceful activity | |
Terminated at step 110 | |
First player won. | |
7[[ 2 2] | |
6 [ 1 ] | |
5 [ 2 1] | |
4 [ ] | |
3 [ -1 2 ] | |
2 [ -2 -2 ] | |
1 [ 1 -2 -1 ] | |
0 [ 1 -2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 10 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Terminated due to peaceful activity | |
Terminated at step 106 | |
Second player won | |
7[[ 2] | |
6 [ 2 1 ] | |
5 [ 2 1 2 ] | |
4 [ ] | |
3 [ -1 -2 ] | |
2 [ -2 -2 ] | |
1 [ -1 -2 -1 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 50 /200 | |
Game step 30 /200 | |
Game step 10 /200 | |
Game step 20 /200 | |
Game step 50 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 60 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 60 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 30 /200 | |
Game step 50 /200 | |
Game step 70 /200 | |
Game step 40 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 40 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 80 /200 | |
Game step 70 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 80 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Terminated due to peaceful activity | |
Terminated at step 85 | |
Second player won | |
7[[ -1 ] | |
6 [ 1 2 -1 ] | |
5 [ 2 1] | |
4 [ 2 1 1 ] | |
3 [ 1] | |
2 [ -2 -1 -2 ] | |
1 [ -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 80 /200 | |
Game step 90 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 90 /200 | |
Terminated due to peaceful activity | |
Terminated at step 98 | |
First player won. | |
7[[ 2 -1 -1] | |
6 [ 2 -1 ] | |
5 [ 2 2 ] | |
4 [ 1 2 1 1 ] | |
3 [ -2 ] | |
2 [ -2 ] | |
1 [ ] | |
0 [ -2 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 90 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 90 /200 | |
Terminated due to peaceful activity | |
Terminated at step 99 | |
First player won. | |
7[[ -1] | |
6 [ 2 1 ] | |
5 [ 2 2 1] | |
4 [ 1 ] | |
3 [ 1] | |
2 [-1 -2 -2 ] | |
1 [ -1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 100 /200 | |
Game step 80 /200 | |
Terminated due to peaceful activity | |
Terminated at step 102 | |
Second player won | |
7[[ -1] | |
6 [ 1 ] | |
5 [ 2 ] | |
4 [ 2 ] | |
3 [ -1 -1 ] | |
2 [ -2 -1 -2 ] | |
1 [ 1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 80 /200 | |
Game step 100 /200 | |
Terminated due to peaceful activity | |
Terminated at step 102 | |
First player won. | |
7[[ -1 -1] | |
6 [ 2 2 1 ] | |
5 [ 1 1 2 1] | |
4 [ 1 1 ] | |
3 [ 1] | |
2 [ -2 -1 -2 ] | |
1 [ -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Terminated due to peaceful activity | |
Terminated at step 87 | |
First player won. | |
7[[ 2 -1 -1] | |
6 [ 2 2 -1 ] | |
5 [ 1 2 ] | |
4 [ 1 1 ] | |
3 [ ] | |
2 [ -2 ] | |
1 [ -1 ] | |
0 [ 1 -2 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 90 /200 | |
Game step 100 /200 | |
Terminated due to peaceful activity | |
Terminated at step 101 | |
First player won. | |
7[[ 2] | |
6 [ 2 ] | |
5 [ 1 2 ] | |
4 [-1 ] | |
3 [ -1 2 -2 ] | |
2 [ -2 ] | |
1 [ 1 -2 -2] | |
0 [ 1 1 ]] | |
0 1 2 3 4 5 6 7 | |
MCTS pool has joined | |
Terminal sentinel is put on queue | |
Sentinel received. GPU will process this batch and terminate afterwards | |
Queue task done signal sent. Queue will join. Thread may still be running. | |
Queue has joined | |
GPU Thread has joined | |
Successful generation of many games? | |
Queue empty: True | |
lowpuct train epoch 8, batch 0. running value loss: 0.03070. running policy loss: 0.06028. running p diff: 0.04063 | |
lowpuct valid epoch 8, batch 0. validation value loss: 0.03211. validation policy loss: 0.05617 validation p diff: 0.03983 | |
saved model lowpuct at D:\Git\alphazero-checker\saves\lowpuct_8_0.pkl | |
lowpuct train epoch 8, batch 10. running value loss: 0.03025. running policy loss: 0.05949. running p diff: 0.04030 | |
lowpuct train epoch 8, batch 20. running value loss: 0.03105. running policy loss: 0.05898. running p diff: 0.04027 | |
lowpuct train epoch 8, batch 30. running value loss: 0.03084. running policy loss: 0.05824. running p diff: 0.03988 | |
lowpuct train epoch 8, batch 40. running value loss: 0.03131. running policy loss: 0.05770. running p diff: 0.03959 | |
lowpuct train epoch 8, batch 50. running value loss: 0.03102. running policy loss: 0.05707. running p diff: 0.03967 | |
lowpuct train epoch 8, batch 60. running value loss: 0.03082. running policy loss: 0.05755. running p diff: 0.04007 | |
lowpuct train epoch 8, batch 70. running value loss: 0.03131. running policy loss: 0.05749. running p diff: 0.04020 | |
lowpuct train epoch 8, batch 80. running value loss: 0.03123. running policy loss: 0.05706. running p diff: 0.04041 | |
lowpuct train epoch 8, batch 90. running value loss: 0.03150. running policy loss: 0.05700. running p diff: 0.04022 | |
lowpuct train epoch 8, batch 100. running value loss: 0.03193. running policy loss: 0.05721. running p diff: 0.04013 | |
lowpuct train epoch 8, batch 110. running value loss: 0.03266. running policy loss: 0.05668. running p diff: 0.04007 | |
lowpuct train epoch 8, batch 120. running value loss: 0.03223. running policy loss: 0.05686. running p diff: 0.04014 | |
lowpuct train epoch 8, batch 130. running value loss: 0.03252. running policy loss: 0.05704. running p diff: 0.04029 | |
lowpuct train epoch 8, batch 140. running value loss: 0.03221. running policy loss: 0.05684. running p diff: 0.04057 | |
lowpuct train epoch 8, batch 150. running value loss: 0.03145. running policy loss: 0.05658. running p diff: 0.04069 | |
lowpuct train epoch 8, batch 160. running value loss: 0.03117. running policy loss: 0.05636. running p diff: 0.04064 | |
lowpuct train epoch 8, batch 170. running value loss: 0.03080. running policy loss: 0.05629. running p diff: 0.04068 | |
lowpuct train epoch 8, batch 180. running value loss: 0.03061. running policy loss: 0.05635. running p diff: 0.04058 | |
lowpuct train epoch 8, batch 190. running value loss: 0.03143. running policy loss: 0.05663. running p diff: 0.04077 | |
lowpuct train epoch 8, batch 200. running value loss: 0.03229. running policy loss: 0.05700. running p diff: 0.04079 | |
lowpuct train epoch 8, batch 210. running value loss: 0.03232. running policy loss: 0.05716. running p diff: 0.04060 | |
lowpuct train epoch 8, batch 220. running value loss: 0.03205. running policy loss: 0.05716. running p diff: 0.04034 | |
lowpuct train epoch 8, batch 230. running value loss: 0.03252. running policy loss: 0.05686. running p diff: 0.04057 | |
lowpuct train epoch 8, batch 240. running value loss: 0.03146. running policy loss: 0.05667. running p diff: 0.04044 | |
lowpuct train epoch 8, batch 250. running value loss: 0.03090. running policy loss: 0.05624. running p diff: 0.04068 | |
lowpuct train epoch 8, batch 260. running value loss: 0.03043. running policy loss: 0.05648. running p diff: 0.04080 | |
lowpuct train epoch 8, batch 270. running value loss: 0.03049. running policy loss: 0.05636. running p diff: 0.04088 | |
lowpuct train epoch 8, batch 280. running value loss: 0.03064. running policy loss: 0.05651. running p diff: 0.04047 | |
lowpuct train epoch 8, batch 290. running value loss: 0.03062. running policy loss: 0.05637. running p diff: 0.04074 | |
lowpuct train epoch 8, batch 300. running value loss: 0.03123. running policy loss: 0.05692. running p diff: 0.04075 | |
lowpuct train epoch 8, batch 310. running value loss: 0.03164. running policy loss: 0.05698. running p diff: 0.04081 | |
lowpuct train epoch 8, batch 320. running value loss: 0.03146. running policy loss: 0.05735. running p diff: 0.04095 | |
lowpuct train epoch 8, batch 330. running value loss: 0.03100. running policy loss: 0.05796. running p diff: 0.04108 | |
lowpuct train epoch 8, batch 340. running value loss: 0.03094. running policy loss: 0.05782. running p diff: 0.04098 | |
lowpuct train epoch 8, batch 350. running value loss: 0.03045. running policy loss: 0.05712. running p diff: 0.04071 | |
lowpuct train epoch 8, batch 360. running value loss: 0.03121. running policy loss: 0.05669. running p diff: 0.04065 | |
lowpuct train epoch 8, batch 370. running value loss: 0.03172. running policy loss: 0.05649. running p diff: 0.04055 | |
lowpuct train epoch 8, batch 380. running value loss: 0.03164. running policy loss: 0.05598. running p diff: 0.04059 | |
lowpuct train epoch 8, batch 390. running value loss: 0.03213. running policy loss: 0.05617. running p diff: 0.04053 | |
lowpuct train epoch 8, batch 400. running value loss: 0.03238. running policy loss: 0.05616. running p diff: 0.04043 | |
lowpuct train epoch 8, batch 410. running value loss: 0.03139. running policy loss: 0.05604. running p diff: 0.04021 | |
lowpuct train epoch 8, batch 420. running value loss: 0.03098. running policy loss: 0.05613. running p diff: 0.03994 | |
lowpuct train epoch 8, batch 430. running value loss: 0.03188. running policy loss: 0.05595. running p diff: 0.04017 | |
lowpuct train epoch 8, batch 440. running value loss: 0.03168. running policy loss: 0.05611. running p diff: 0.03997 | |
lowpuct train epoch 8, batch 450. running value loss: 0.03191. running policy loss: 0.05625. running p diff: 0.03954 | |
lowpuct train epoch 8, batch 460. running value loss: 0.03229. running policy loss: 0.05691. running p diff: 0.03991 | |
lowpuct train epoch 8, batch 470. running value loss: 0.03232. running policy loss: 0.05688. running p diff: 0.04016 | |
lowpuct train epoch 8, batch 480. running value loss: 0.03133. running policy loss: 0.05699. running p diff: 0.04041 | |
lowpuct train epoch 8, batch 490. running value loss: 0.03094. running policy loss: 0.05675. running p diff: 0.04059 | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Terminated due to peaceful activity | |
Terminated at step 77 | |
Second player won | |
7[[ 2] | |
6 [ 1 2 1 ] | |
5 [ 2 1] | |
4 [ ] | |
3 [ ] | |
2 [ -2 -1 ] | |
1 [ -1 -2 ] | |
0 [-2 -2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 70 /200 | |
Game step 80 /200 | |
Terminated due to peaceful activity | |
Terminated at step 75 | |
Second player won | |
7[[ 2 -1] | |
6 [ 2 1 ] | |
5 [ 2 1] | |
4 [ 1 ] | |
3 [ -1 ] | |
2 [-1 -2 ] | |
1 [ -1 -2 -2] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 10 /200 | |
Game step 80 /200 | |
Game step 90 /200 | |
Game step 10 /200 | |
Game step 80 /200 | |
Game step 90 /200 | |
Terminated due to peaceful activity | |
Terminated at step 95 | |
First player won. | |
7[[ 2] | |
6 [ 2 2 1 ] | |
5 [ 1 2 ] | |
4 [ 1 ] | |
3 [ ] | |
2 [-1 -2 -2 ] | |
1 [ -1 -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 90 /200 | |
Game step 90 /200 | |
Game step 90 /200 | |
Game step 10 /200 | |
Terminated due to peaceful activity | |
Terminated at step 94 | |
First player won. | |
7[[ 2] | |
6 [ 2 ] | |
5 [ 2 2 ] | |
4 [ -1 ] | |
3 [ -1 1 1 ] | |
2 [ -2 -2 ] | |
1 [ 1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 30 /200 | |
Game step 90 /200 | |
Game step 100 /200 | |
Game step 30 /200 | |
Terminated due to peaceful activity | |
Terminated at step 103 | |
Second player won | |
7[[ -1] | |
6 [ 2 1 ] | |
5 [ 2 2 1] | |
4 [ 1 ] | |
3 [ -1 ] | |
2 [-1 -2 -2 ] | |
1 [ -1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 100 /200 | |
Game step 100 /200 | |
Game step 10 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Terminated at step 51 | |
First player won. | |
7[[ 2] | |
6 [ 1 ] | |
5 [ 1 2 ] | |
4 [ 1 ] | |
3 [ ] | |
2 [-1 2 1 ] | |
1 [ 1 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 10 /200 | |
Game step 100 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 110 /200 | |
Game step 10 /200 | |
Game step 110 /200 | |
Game step 50 /200 | |
Terminated due to peaceful activity | |
Terminated at step 111 | |
Second player won | |
7[[ 2 -1] | |
6 [ 1 ] | |
5 [ 2 ] | |
4 [ 1 -2 ] | |
3 [ 2 -2 ] | |
2 [-1 -2 ] | |
1 [ 1 -2 -1 ] | |
0 [ 1 -2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Terminated due to peaceful activity | |
Terminated at step 108 | |
First player won. | |
7[[ -1] | |
6 [ 2 2 1 ] | |
5 [ 2 1 2 ] | |
4 [ 1 1 ] | |
3 [ 1 -2 ] | |
2 [ -2 -2 ] | |
1 [ -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 120 /200 | |
Game step 40 /200 | |
Game step 10 /200 | |
Game step 30 /200 | |
Game step 60 /200 | |
Game step 10 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 130 /200 | |
Game step 30 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Terminated at step 22 | |
Second player won | |
7[[ -1] | |
6 [-1 -1 -1 ] | |
5 [ -1 -1 -1 -1] | |
4 [-1 -1 -1 -1 ] | |
3 [ 1 1 1 1] | |
2 [ 1 1 1 1 ] | |
1 [ 1 1 1 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 50 /200 | |
Game step 70 /200 | |
Terminated due to peaceful activity | |
Terminated at step 134 | |
First player won. | |
7[[ 2] | |
6 [ 2 1 ] | |
5 [ 2 2 ] | |
4 [ -2 ] | |
3 [ 2 -2 ] | |
2 [ -2 -2 ] | |
1 [ -1 -2 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 30 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 40 /200 | |
Game step 60 /200 | |
Game step 40 /200 | |
Game step 80 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 70 /200 | |
Game step 90 /200 | |
Game step 70 /200 | |
Terminated due to peaceful activity | |
Terminated at step 90 | |
First player won. | |
7[[ -1] | |
6 [ 1 ] | |
5 [ 1 2 1] | |
4 [ 2 1 ] | |
3 [ ] | |
2 [-1 -2 -1 ] | |
1 [ -1 -2 ] | |
0 [ 1 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 70 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 90 /200 | |
Game step 90 /200 | |
Terminated due to peaceful activity | |
Terminated at step 95 | |
Second player won | |
7[[ -1] | |
6 [ 2 1 ] | |
5 [ 2 ] | |
4 [ ] | |
3 [ -2 ] | |
2 [-1 -2 -1 ] | |
1 [ -1 -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 90 /200 | |
Game step 80 /200 | |
Terminated due to peaceful activity | |
Terminated at step 90 | |
First player won. | |
7[[ 2 -1 -1] | |
6 [ 2 2 -1 ] | |
5 [ 2 -2 2 ] | |
4 [ 1 ] | |
3 [ -1 1] | |
2 [-1 -2 ] | |
1 [ -1 ] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 80 /200 | |
Terminated due to peaceful activity | |
Terminated at step 89 | |
Second player won | |
7[[ 2] | |
6 [ 2 1 ] | |
5 [ 2 1] | |
4 [ 2 ] | |
3 [ -1 -2 ] | |
2 [-1 -2 ] | |
1 [ -1 -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 100 /200 | |
Terminated due to peaceful activity | |
Terminated at step 104 | |
Second player won | |
7[[ 2 -1 ] | |
6 [ 1 2 -1 ] | |
5 [ 2 1] | |
4 [ 2 -2 ] | |
3 [ ] | |
2 [ -2 -2 ] | |
1 [ -2 ] | |
0 [-2 ]] | |
0 1 2 3 4 5 6 7 | |
Game step 90 /200 | |
Game step 100 /200 | |
Terminated due to peaceful activity | |
Terminated at step 104 | |
First player won. | |
7[[ 2 -1 -1] | |
6 [ 2 2 -1 ] | |
5 [ 2 1 2 ] | |
4 [ 2 -2 1 ] | |
3 [ ] | |
2 [ -2 ] | |
1 [ -2 ] | |
0 [ ]] | |
0 1 2 3 4 5 6 7 | |
MCTS pool has joined | |
Terminal sentinel is put on queue | |
Sentinel received. GPU will process this batch and terminate afterwards | |
Queue task done signal sent. Queue will join. Thread may still be running. | |
Queue has joined | |
GPU Thread has joined | |
Successful generation of many games? | |
Queue empty: True | |
lowpuct train epoch 9, batch 0. running value loss: 0.03084. running policy loss: 0.05677. running p diff: 0.04116 | |
lowpuct valid epoch 9, batch 0. validation value loss: 0.03175. validation policy loss: 0.05418 validation p diff: 0.03996 | |
saved model lowpuct at D:\Git\alphazero-checker\saves\lowpuct_9_0.pkl | |
lowpuct train epoch 9, batch 10. running value loss: 0.03080. running policy loss: 0.05640. running p diff: 0.04120 | |
lowpuct train epoch 9, batch 20. running value loss: 0.03119. running policy loss: 0.05609. running p diff: 0.04132 | |
lowpuct train epoch 9, batch 30. running value loss: 0.03193. running policy loss: 0.05580. running p diff: 0.04126 | |
lowpuct train epoch 9, batch 40. running value loss: 0.03244. running policy loss: 0.05567. running p diff: 0.04086 | |
lowpuct train epoch 9, batch 50. running value loss: 0.03215. running policy loss: 0.05575. running p diff: 0.04033 | |
lowpuct train epoch 9, batch 60. running value loss: 0.03176. running policy loss: 0.05571. running p diff: 0.03987 | |
lowpuct train epoch 9, batch 70. running value loss: 0.03150. running policy loss: 0.05556. running p diff: 0.03954 | |
lowpuct train epoch 9, batch 80. running value loss: 0.03111. running policy loss: 0.05542. running p diff: 0.03918 | |
lowpuct train epoch 9, batch 90. running value loss: 0.03119. running policy loss: 0.05507. running p diff: 0.03965 | |
lowpuct train epoch 9, batch 100. running value loss: 0.03133. running policy loss: 0.05482. running p diff: 0.04020 | |
lowpuct train epoch 9, batch 110. running value loss: 0.03180. running policy loss: 0.05476. running p diff: 0.04023 | |
lowpuct train epoch 9, batch 120. running value loss: 0.03182. running policy loss: 0.05471. running p diff: 0.04054 | |
lowpuct train epoch 9, batch 130. running value loss: 0.03156. running policy loss: 0.05446. running p diff: 0.04053 | |
lowpuct train epoch 9, batch 140. running value loss: 0.03146. running policy loss: 0.05432. running p diff: 0.04013 | |
lowpuct train epoch 9, batch 150. running value loss: 0.03150. running policy loss: 0.05419. running p diff: 0.04011 | |
lowpuct train epoch 9, batch 160. running value loss: 0.03090. running policy loss: 0.05421. running p diff: 0.04018 | |
lowpuct train epoch 9, batch 170. running value loss: 0.03098. running policy loss: 0.05410. running p diff: 0.04007 | |
lowpuct train epoch 9, batch 180. running value loss: 0.03166. running policy loss: 0.05440. running p diff: 0.04003 | |
lowpuct train epoch 9, batch 190. running value loss: 0.03123. running policy loss: 0.05461. running p diff: 0.04033 | |
lowpuct train epoch 9, batch 200. running value loss: 0.03092. running policy loss: 0.05499. running p diff: 0.04040 | |
lowpuct train epoch 9, batch 210. running value loss: 0.03104. running policy loss: 0.05443. running p diff: 0.04068 | |
lowpuct train epoch 9, batch 220. running value loss: 0.03082. running policy loss: 0.05473. running p diff: 0.04082 | |
lowpuct train epoch 9, batch 230. running value loss: 0.02996. running policy loss: 0.05470. running p diff: 0.04034 | |
lowpuct train epoch 9, batch 240. running value loss: 0.03057. running policy loss: 0.05484. running p diff: 0.04023 | |
lowpuct train epoch 9, batch 250. running value loss: 0.03045. running policy loss: 0.05445. running p diff: 0.04023 | |
lowpuct train epoch 9, batch 260. running value loss: 0.03082. running policy loss: 0.05473. running p diff: 0.04023 | |
lowpuct train epoch 9, batch 270. running value loss: 0.03139. running policy loss: 0.05417. running p diff: 0.04029 | |
lowpuct train epoch 9, batch 280. running value loss: 0.03170. running policy loss: 0.05405. running p diff: 0.04083 | |
lowpuct train epoch 9, batch 290. running value loss: 0.03135. running policy loss: 0.05410. running p diff: 0.04094 | |
lowpuct train epoch 9, batch 300. running value loss: 0.03119. running policy loss: 0.05402. running p diff: 0.04093 | |
lowpuct train epoch 9, batch 310. running value loss: 0.03137. running policy loss: 0.05408. running p diff: 0.04110 | |
lowpuct train epoch 9, batch 320. running value loss: 0.03092. running policy loss: 0.05444. running p diff: 0.04087 | |
lowpuct train epoch 9, batch 330. running value loss: 0.03055. running policy loss: 0.05451. running p diff: 0.04065 | |
lowpuct train epoch 9, batch 340. running value loss: 0.03092. running policy loss: 0.05392. running p diff: 0.04064 | |
lowpuct train epoch 9, batch 350. running value loss: 0.03139. running policy loss: 0.05401. running p diff: 0.04036 | |
lowpuct train epoch 9, batch 360. running value loss: 0.03105. running policy loss: 0.05428. running p diff: 0.04032 | |
lowpuct train epoch 9, batch 370. running value loss: 0.03148. running policy loss: 0.05422. running p diff: 0.04042 | |
lowpuct train epoch 9, batch 380. running value loss: 0.03131. running policy loss: 0.05422. running p diff: 0.04068 | |
lowpuct train epoch 9, batch 390. running value loss: 0.03088. running policy loss: 0.05429. running p diff: 0.04071 | |
lowpuct train epoch 9, batch 400. running value loss: 0.03053. running policy loss: 0.05411. running p diff: 0.04090 | |
lowpuct train epoch 9, batch 410. running value loss: 0.03043. running policy loss: 0.05367. running p diff: 0.04088 | |
lowpuct train epoch 9, batch 420. running value loss: 0.02994. running policy loss: 0.05354. running p diff: 0.04090 | |
lowpuct train epoch 9, batch 430. running value loss: 0.03066. running policy loss: 0.05340. running p diff: 0.04097 | |
lowpuct train epoch 9, batch 440. running value loss: 0.03113. running policy loss: 0.05380. running p diff: 0.04112 | |
lowpuct train epoch 9, batch 450. running value loss: 0.03127. running policy loss: 0.05412. running p diff: 0.04096 | |
lowpuct train epoch 9, batch 460. running value loss: 0.03164. running policy loss: 0.05437. running p diff: 0.04091 | |
lowpuct train epoch 9, batch 470. running value loss: 0.03141. running policy loss: 0.05418. running p diff: 0.04091 | |
lowpuct train epoch 9, batch 480. running value loss: 0.03115. running policy loss: 0.05431. running p diff: 0.04117 | |
lowpuct train epoch 9, batch 490. running value loss: 0.03006. running policy loss: 0.05442. running p diff: 0.04090 | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 0 /200 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 0 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 20 /200 | |
Terminated at step 22 | |
Second player won | |
7[[ -1 ] | |
6 [-1 -1 -1 ] | |
5 [ -1 -1 -1 -1] | |
4 [-1 -1 -1 -1 ] | |
3 [ 1 1 1 1] | |
2 [ 1 1 1 1 ] | |
1 [ 1 1 1] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Terminated at step 22 | |
Second player won | |
7[[ -1] | |
6 [-1 -1 -1 ] | |
5 [ -1 -1 -1 -1] | |
4 [-1 -1 -1 -1 ] | |
3 [ 1 1 1 1] | |
2 [ 1 1 1 1 ] | |
1 [ 1 1 1] | |
0 [ 1 ]] | |
0 1 2 3 4 5 6 7 | |
Generating a new game with MCTS | |
Game step 0 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 20 /200 | |
Game step 10 /200 | |
Game step 10 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 40 /200 | |
Game step 30 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 20 /200 | |
Game step 30 /200 | |
Game step 30 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 30 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 60 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 40 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 50 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 60 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 80 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 70 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 70 /200 | |
Game step 80 /200 | |
Game step 80 /200 | |
Game step 90 /200 | |
Game step 80 /200 |
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment