OneDayItllWork wrote:
More theoretical stuff here I'm afraid.
This answer is theoretical too since I don't really know
OneDayItllWork wrote:
If we've got some optimal convergence algorithm that uses chance based sampling and has multiple bet amounts as possible actions, is that likely to skew the bot towards being too aggressive due to variance? In my head, it is, if not, why not? Let's say we have our possible actions as Fold, Check/Call and then 100000 different bet sizes (extreme example here), due to variance, on marginal hands the chances are one of of those bet actions will skew towards being more +EV than the passive actions as there are so many of them.
Are you saying that for a particular bet amount there are too few iterations to achieve convergence?
OneDayItllWork wrote:
If this is the case, can we measure it?
There is surely a relationship between sample size, variance and error. Maybe you could get a estimate of that from examining the history of convergence of more frequently visited nodes.
OneDayItllWork wrote:
What can we do about it?
- increase the number of iterations
- reduce the number of bet actions
- concentrate the iterations on the area of interest
- smooth the results over "nearby" points using some sort of machine learning. Depending on your abstraction, "nearby" might be hard to define. Since performance is likely to be an issue maybe something like this
http://webdocs.cs.ualberta.ca/~sutton/b ... ode88.html