-
Notifications
You must be signed in to change notification settings - Fork 0
/
net_notes.txt
40 lines (34 loc) · 1.52 KB
/
net_notes.txt
1
2
3
4
5
6
7
8
9
10
11
12
13
14
15
16
17
18
19
20
21
22
23
24
25
26
27
28
29
30
31
32
33
34
35
36
37
38
39
40
** Feb 28, 2012
With 15,000 games of self-play with a neural network, got results:
SimpleAI: 38.2% +/- 4.0%
Tie: 5.4% +/- 1.8%
BigMoneyUltimate: 56.4% +/- 4.0%
This roughly matches the best performance achieved with the previous
technique of training on BMU self-play using batch learning. These
weights were however achieved by self-play and incremental learning,
which hopefully means there is more room for improvement.
I am tracking a new statistic that hopefully can help with tweaking
the learning rate: number of mistakes, which are
almost-certainly-incorrect moves such as buying a copper over a silver.
With some tweaking to the learning rate based on mistake rate, 20,000
games yielded:
SimpleAI: 41.2% +/- 0.9%
Tie: 6.3% +/- 0.4%
BigMoneyUltimate: 52.5% +/- 0.9%
Not too bad.
Did another 20,000:
SimpleAI: 40.4% +/- 0.9%
Tie: 6.8% +/- 0.5%
BigMoneyUltimate: 52.8% +/- 0.9%
Looking at the mistake graph, it seems like the initially-high learning
rate is causing a massive spike in mistakes (nearly to 50%!) in the
first ~3000 games. Since that is 15% of games where SimpleAI is basically
throwing them away, it seems worth trying a non-learning run to see its
"true" strength.
SimpleAI: 41.8% +/- 0.9%
Tie: 6.6% +/- 0.5%
BigMoneyUltimate: 51.6% +/- 0.9%
This is barely different. Surprising I think. The mistake graph still shows
the same shape--so maybe there is a problem with how we are measuring mistakes.
Oh shit yes, I was still reading the old file. Also, it seems like 100 might
be too small a value for ROLLING.