jukofyork wrote:
When you say "rollout" do you mean:
1) You are doing just a single rollout of the remaining cards and allocating the pot accordingly.
2) You are doing several monte-carlo rollouts of the remaining cards and averaging to get a final EV/equity expectation to allocate.
3) You traversing the tree of all possible remaining rollouts and averaging to get a final EV/equity expectation to allocate.
I'm doing 1) if I understand you right. So on all-in on flop or turn, I rollout all remaining cards. I enumerate all possibilities, (remainingCards choose 1) or (remainingCards choose 2) evaluations. I get basically no preflop all-in with my bots (except on AA vs AA). So that I handle that like all-in on flop should not be a problem (get the regular flop and rollout the rest).
I don't see how doing 2) or 3) gives different results. Wouldn't the 2) and 3) approximate what I get with 1)? Except for preflop all-in of course.
Or do you mean not just rolling out on all-in, but for not-all-in too? I cannot give the bots all possible river cards and see what they would do. Decisions based on different cards are expensive in my approach (because I have to cache calculations for each card).
jukofyork wrote:
It sounds like your evolutionary algorithm is exploiting the fact that choosing to go all-in often maximizes variance, and then this in turn gives these members of the population a good chance of coming out on top due to random sampling?
Not sure I get that. With rollout the all-in minimizes the variance. Because they no longer have the variance of the remaining cards. So yes, for the bots without rollout, they see more variance in the all-in case compared to the rollout bots during training. But I don't see how they can exploit that, if I compare the 2 approaches I let the bots run on a other set of hands, so the larger variance could equally likely bite them in the ass. So intuitively I would say the rollout bots should have an advantage if I use rollout on comparing the 2, because they expect the lower variance they will see? But they did not seem to (however, as said earlier the tests don't say much).
It is a good point. I probably need to get many more validation hand sets to see how large the variance for fights between the 2 approaches is.
@all: thanks guys. I see I need more detailed data. I think I need more validation hand sets to measure the variance for fights between the 2 approaches. And I probably need a larger fixed hand pool used for training. And I should probably check the effect of the very few preflop all-in.
I probably have to train bots with both approaches for some weeks to do a better comparison. Even then could still be that one approach just luckily found a better local minima Statistics: Posted by SkyBot — Sun Jul 05, 2015 1:40 pm
]]>