I'm not sure if I completely understand what you're wanting to achieve, but I think you're trying to go from a huge model that maps game states to action probabilities and then model this information in a NN. Therefore decreasing the size ~250Gb to something far, far smaller, am I close?
Nasher wrote:
I tried breaking up the games into stacks <> 20bb, but it didn't seem to make a difference. The whole idea of regressing multiple stack sizes is to allow the NN to self-interpolate between them. Same goes for the various game states.
Have you tried any other techniques for breaking up the games, even more? It sounds like there is
a lot of information you're trying to encode into the NN. I feel like the only way you could get a decent model is by splitting the network into, quite a few, smaller models, e.g. have a network that looks at different stacks, split for each stage of the game, split into how many players are active, split into... etc...
You've may have already thought of/tried doing so, or maybe i'm completely off track?
Even for going the opposite way around, i.e. building NNs based on models, I feel like this could be the only way of really getting accurate results.