A couple of months ago I was advised that a good sampling method to try was to divide all of the games in my sample into thirds across all dates. All the games would be sorted chronologically and then numbered 1, 2, 3, 1, 2, 3, etc. and I would then select one of the sets as my test sample. It was proposed that by doing this I would eliminate any seasonal bias. I could then do as much number crunching as I wanted on one of the sets and then test these assumptions on the other 2 sets which were completely out of sample. Although this seemed logical to me, I had been getting some strange results from the verification sets. For example, if I had a "system" that had an ROI of +20% in my initial set, I might get an ROI of -6% and -8% on each of the other 2 sets. Although, the ROI of the entire sample above was +2.66%, I would get negative results in my test sample.
If I do the same thing but divide the 3 sets chronologically (i.e. Set 1=2004 Season, Set 2=2005 Season, Set 3=2006 Season), when my initial set was positive, I would more times than not get positive results in the 2 out of sample test sets.
Empirically, for reasons I still do not understand, this peculiar effect has lead me to believe that the prediction of sports results is closer to a card game (say blackjack for instance) then it is to dice.
Here is my point:
Let's say you gave me 10,000 game betting results in date order from any major sport and I selected every second game for a total of 5,000. Purely by chance the favoured team in this test sample won 100% of the time for an ROI of +50% betting to win 1 unit. Now, if sports betting was like dice which have no memory, I would expect that the ROI of favourites in the 5,000 games that were out of sample would be close to -vig or -2% at a 5 cent book. However, I would bet the farm that the ROI of underdogs would be VERY high in the out of sample set.
My conclusion:
Although I do not have the math skills to prove this, based on my recent experience I would suggest at least one of the following is true:
a) If data is divided across dates in an attempt to remove seasonal bias, the results of this original sample and the out of sample data need to be totaled together to produce a final result.
OR
b) Data should not be divided across dates and should be divided by logical dates (i.e. by whole seasons and not something like using pre-season game data to predict playoff game results etc.)
I highly respect the person that gave me the original suggestion but I am simply not able to produce consistent results by this sampling method. I would appreciate any thoughts, especially those that involve the math of what I am saying.