1. #1
    Ganchrow
    Nolite te bastardes carborundorum.
    Ganchrow's Avatar Become A Pro!
    Join Date: 08-28-05
    Posts: 5,011
    Betpoints: 1088

    Market Efficiency and Bayesian Probability Estimation via the Beta Distribution

    Let's start off by making a very general observation about betting markets: Betting markets are inherently competitive -- they reward winners and penalize losers. This creates a dynamic fundamentally different from what we see in, say, internet polling where there's no self-reinforcing drive to succeed.

    Outside the betting world and without money on the line there's little incentive for accuracy.

    Within the betting world, however, people make their opinions known with a signal more telling than their vote -- their money. Within the competitive marketplace it's democracy be damned and capitalism rules the roost.

    (Now certainly this signal can get corrupted by such forces as recreational bettors who'll blindly bet their home team at any odds or by games thrown by dishonest referees or players. Market economists account for these factors by assuming that the smart money (including both books and players betting for profit and with superior information flow and processing), will in totality eventually overwhelm the rest.)

    This leads to a framework known as (profit-wise) "market efficiency".

    Conceptually, the way we'll be considering what market efficiency as applied to sports betting says is that without "extraordinary knowledge" (including meta-knowledge) a player's expected profit from betting at the market no-vig line is on average zero.

    This definition still allows for many interpretations of market efficiency, running the continuum from so-called "strong form" (most restrictive in its assumptions as to market behavior) to "weak form" (least restrictive). But at its most fundamental the defining parameter comes down to little more than what exactly is meant by the characterization of knowledge as "extraordinary".

    Following is an indicative list of examples of types of knowledge going from those likely deemed extraordinary by all but adherents to the strongest forms of efficiency to those only so deemed (possibly) by proponents of the weakest:
    1. first-hand knowledge of game fixing
    2. insider information regarding a change in directive for officials
    3. insider information regarding a key player's health
    4. faster receipt and processing of new information
    5. faster response time following receipt of new information
    6. access to non-readily available (but still public) information
    7. novel quantitative processing of existing information
    8. any proprietary processing of existing information
    9. information gained by watching ESPN
    10. deeper fundamental understanding of motivations and factor facing one's home team or University
    Some strong form advocates would say that while they wouldn't in theory doubt the efficacy of some of the items on the above list, they would call in to question their general existence (or might just point to the infrequency with which such knowledge is found and so tend to consider them little more than just pathological cases).

    Personally, I'll just leave my own characterization as purposely vague with the eventual intention of providing some form of weak parameterization of the concept. I will say that given the above listing I'd probably be somewhere between the middle and the bottom.

    Another question one should ask would be what exactly represents a market line? Here there's a lot less debate. A market line should be one accessible (and already widely bet) to the betting public as a whole and one that won't be artificially held in place by a book limiting the number of bets by individual players (in other words after a reasonable period of time a market line should either move out of the way or accept another bet at the same price).

    The best examples for many sports would probably be Pinnacle, BetFair, and Matchbook. For certain geographically disparate sports, however, one would probably tend to find a better indication of the true market line with some of the Asian books. Generally speaking, very high limits, low juice, market prominence, and attitude towards professionals will jointly determine how well a given sports book's lines are indicative of the market as a whole.

    For the sake of argument we'll just say "the market" is represented by the inside line across geographically-appropriate "Pinnacle-like" books and exchanges. (Within what markets and under what circumstances Pinnacle itself would actually represent a "Pinnacle-like" book would be very much up for debate.)

    Getting right down to it, recall that we that declared the defining characteristic of market efficiency to be that a non-extraordinary player's expected profit from betting at the market no-vig line would on average be zero (or equivalently that on average E[player profit] = -theoretical hold, regardless of market side chosen).

    Recall that for a true win probability of p and a no-vig decimal line of d:

    E[profit] = p*d - 1 = 0
    d = 1 p
    But this definition of market odds is intuitively rather unsatisfying as it presupposes exact knowledge of win probability. Remember that the only defining characteristic of market efficiency to which we've limited ourselves is that (absent extraordinary information, which we discount for the time being) expected profit on bets placed at the zero-vig line will always be zero. But does this really require that the market knows the true win probability exactly all the time every time without margin for error?

    What we're trying to get at here are some determinations about how the market might set its lines (and although we keep referring to "the market" as a single entity, understand that even though the market line might be reflected at only a single book, it the market as a whole represents a conglomeration of opinions across professional players and books. The mathematical mechanics of how "the market" might handle such a conglomeration this we won’t delve into.

    I expect that anyone who's already gotten this far is already familiar with the concept of expectation as it relates to, say, value. Specifically, the expected value of a bet could be described as the theoretical average profit per game were the game to be repeated an infinite number of times (at least that's how a "frequentist" would describe it).

    So we'll write:
    E[p̂] = 1 d
    d = 1 E[p̂]
    Where the caret-looking symbol lurking over (or nearly over) the p is known by the highly complex term "hat". In probability, a hat-ed variable almost always represents an estimate of a distribution parameter.

    But how exactly should we interpret the seemingly innocuous term "E[p̂]"? Well all we're doing is applying the concept of expectations to a random variable.

    It's just that in this case that random variable happens not to be an estimate of value but rather of probability ... more specifically the market's estimate of the true win probability of the game.

    Think about it in a manner similar to expected value. If we repeated the game an infinite number of times and then repeated that set of infinite repetitions an infinite number of times what we'd see is that across each one of those infinite sized game-groups, the frequency of wins was on average the same as the expected win probability (note that this abstraction to an infinite number of infinitely sized game groupings isn't necessary and is only presented with the hope that it might aid in visualization).

    This is an extremely important concept. We're not forcing the assumption that the market is always correct in its probability estimates ... just that its probability estimates are correct on average.

    And there's another way to consolidate the above:
    E[p̂] = p
    So in other words, the expected value of the market estimate of probability equals the true probability. In more technical terms this statement can be read to say that p̂ is an unbiased estimator of p. So while a given probability estimate might be wrong in any single case it's still correct on average. This is perfectly analogous to expected value.

    (FWIW, note that this is not an equivalent statement to the similar E[1/p̂] = 1/p, so in other words while we’ve concluded based on our assumptions that the zero-vig implied probability is an unbiased estimator of true probability, we can't in general conclude that the zero-vig line will be an unbiased estimator of the true line.)

    So now that we've looked at the expected value of the market probability estimate, the next logical meta-statistic would be the variance of the market probability estimate. Once again we're not deviating from well-trod territory -- this is analogous to not only talking about the expected value of a bet, but also talking about the variance of the value of a bet.

    For a reason that should soon become I'm going to switch over to time subscripts and propose a very simple model of market efficiency such that the market estimate of win probability at ordinal time t, p̂t (where t = 0, represents the present and t = Τ represents the market close) obeys:
    1. E(p̂t) = p ∀ 0 ≤ t ≤ Τ
    2. Var(p̂s) > Var(p̂t) > 0 ∀ s > t ≥ Τ

    So what this is saying is that at any moment in time, the reciprocal of the market line (which is to say, p̂t) is an unbiased estimator of true win probability and that the variance of that estimate is strictly decreasing as game time approaches.

    As a corollary to 1) above we can write:
    E(p̂t) = E(p̂0) ∀ 0 ≤ t
    Or in other words the expectation of the implied market probability at any point in the future is just the current implied market probability. This follows directly from the unbiasedness of the probability estimator across time.

    Let's step back and consider this all for a moment. We're saying that at any point in time between now and game start the market estimate of probability will on average be correct and will become continuously more accurate (meaning less variance) as time progresses.

    A good way to intuit this would be in terms of weather forecasting, say the probability of precipitation on a given day. In theory we might say that every forecast we hear we believe to be unbiased (correct on average) and that we further expect the reliability of each forecast to increase (lower variance) as we get closer to the day for which we're forecasting.

    Moving back to statement 1) above, let's try to find some sort of additional functional bound on the market estimator. One particularly apparent bound should be fairly easy to spot. If we're estimating some probability p, and know that that estimate will be correct on average what’s the noisiest possible probability estimate that will still be unbiased?

    It should eventually become apparent that the maximum variance estimator p̂, would simply be the estimator that takes on the value p̂=100% with probability=p, and the value of p̂=0% with probability=(1-p).

    This is just another way of saying that p̂ were Bernoulli distributed with parameter p. (Once again think of weather forecasting ... a Bernoulli forecast of this type would always offer either a 100% or 0% probability of rain on a given future day, and while it might be correct on average -- and remember that the true probability of rain on a future is never really 0% or 100% -- it'll certainly be less valuable in picking a future day for a picnic than a lower variance probability estimate).

    We know that a Bernoulli-distributed variable with parameter probability parameter p has variance = p*(1-p), so when taken in conjunction with condition 2) which says that future variance of a given estimate will always be lower than the current variance, this gives another bound on Var(p̂2). Specifically, 2) implies that maximum variance of p̂t could only be achieved at t=&-infin;, which means that for all finite t:

    Var(p̂t) < p*(1-p)
    Recalling that for any random variable X:
    Var(X) = E[X2] - E[X]2
    We could also say:

    E[p̂2] < p*(1-p) + p2 ≡ p

    These are by necessity rather weak statements, but they nevertheless do give lend a bit more probabilistic rigor to the notion of market efficiency by allowing us to place functional bounds on the estimator p̂:

    E(p̂) = p
    Var(p̂) < p*(1-p)

    But this upper bound on Var(p̂) is the result of assuming a maximum variance estimator of p, which was that obtained by the Bernoulli distribution and a high variance estimator like just doesn't pass the smell test. A probability of 0%/100% would yield a bet at +∞ on one team or the other. Whatever assumptions we might accept that's just a full out absurd result. So we'd really hope that we can improve upon that.

    But where to turn? Given that the Beta distribution is the conjugate prior of the binomial it certainly might be convenient to look there.

    Recalling our vast knowledge of probability distributions (or Wikipedia) we can say that if p̂ is Beta distributed with mean p and Beta shape parameter ν, then the variance of p̂ is given by:

    1. Var(p̂) = p*(1-p) 1+ν


    It's via this parameter ν (the Greek letter nu) that we can start thinking of providing a parameterization of the concept of market efficiency.

    Specifically ν → 0, represents the maximum variance unbiased estimator of p, while ν → ∞ would yield constant win probability p̂ ≡ p. And anywhere else in between is fair game.

    This is actually a meaningful result. What we've done is use a parameter from a well-known distribution to place an equality bound on the variance of the market probability estimate effectively parameterizing the degree of efficiency for a given market. Of course we're still assuming a magically unbiased line that's even more magically taking into account all market knowledge ... but patience is a virtue.

    To continue further, we’ll consider our model from the perspective of Bayesian inference.

    What is Bayesian inference you might ask?

    Well it's really a different way of looking at probability that allows for a different methodology in forecasting. Recall that I earlier wrote that a frequentist would view the expected value of a bet as the average profit per game were the bet to be repeated an infinite number of times. Well that's just not how a Bayesian sees the world.

    A Bayesian considers the probability of an outcome as a (possibly subjective) measure of the degree of informed belief in that outcome. We talk about "informed" belief because the process of Bayesian inference involves updating one's prior beliefs based on the availability of new evidence.

    A Bayesian doesn't in general think about hypothetical frequencies of an event given a hypothetical infinite number of repetitions because in general events can't be repeated an infinite number. To estimate outcome probability what a Bayesian does is gauge prior knowledge of that event and then update that knowledge as future information becomes available.

    Note that this methodology doesn't hold value for all types of experiments as for some events can know everything there is to know about it a priori. These events (take the game of Craps for instance) are frequentist in nature meaning that there is no value to new information (although Craps could deemed be otherwise were we to suspect either an unfair game or the presence of a skilled dice roller).

    This can be a particular convenient tool when looking at the progression of a betting line. One can build a forecast in any way one chooses and then continually reevaluate that forecast based on upon the availability of new evidence (e.g., a change in the market line).

    By way of contrast, a non-Bayesian might place a bet at a line of +3, and then after observing the line move to +5 simply declare his original bet "bad" in that the market hadn't backed up his opinion. A Bayesian, on the other hand would realize that his bet was made conditioned only on the information he had available at the time the bet was made (which would have only included the then current line), and while he would almost certainly view the bet as "unfortunate", he could accept that the bet was still a good bet at the time it was placed. Certainly he'd use the new information to revise his current opinions on game probabilities, and going a step further might even use it to discount the value of his model, but a Bayesian can accept that a decision might be perfectly valid at the time it was made, even as new information sheds doubt on it in hindsight.

    At the heart of all Bayesian inference is what's known as Bayes' theorem. This is stated as follows:

    Let H = A hypothesis (for example the hypothesis that win probability = a specific value p̂)
    Let E = The evidence

    Prob(H|E) = P(E|H) * P(H) / P(E)
    What this says is that the probability of a hypothesis (which we can extend to a probability distribution) given some set of evidence is proportional to how likely we'd be to view that evidence assuming our hypothesis true times the probability of the hypothesis based solely on prior information (i.e., without benefit of any new evidence).

    Armed with this added understanding, we can meaningfully back off from our prior view of market efficiency by relaxing the global assumption of p̂ being a strictly unbiased estimator of p and replace it with the weaker assumption:

    p̂(K) = E(p | Κ)

    where Κ represents a given "knowledge space" comprised not only of timely and historical event and market data but also of specific prior knowledge of relevant deterministic functional relationships between that data and p̂, the variable of interest.

    Note a couple of things about this equation.

    Firstly, we've moved our p̂ estimator outside of the expectation and moved the parameter p representing "true" market probability inside the expectation. This is consistent with a Bayesian framework. The actual probability value is itself viewed as a random variable, while the estimator of p is viewed as a dependant variable. This is because Bayesian true probability isn't viewed as a discrete measurable quantity and can only viewed in terms of information upon which estimates are conditioned.

    Secondly, note that we're now making explicit the functional relationship between our estimator of p̂ and the knowledge space. This reflects the notion that the estimator is only as valid as the knowledge upon which it's based. Also included in K would ν(K), which once again represents a measure of inherent "value" of the knowledge space, only in this case relating that value not only to the variance of the estimate but also to its mean.

    This is virtually akin to stating that p̂ is a minimum mean square error (MMSE) estimator of p conditioned on "prior knowledge". It isn't exactly the same because we're necessitating explicit inclusion of "relevant deterministic functional relationships" within the space. A true MMSE estimator would need to already take into account all possible functional relationships between information pieces.

    Still what this allows for is an explicit decomposition of "market probability" into any number of discrete conditionally independent knowledge packets. What’s the relevance of this? Well it’s actually quite substantial. Remember our vague "extraordinary knowledge" constraint? Well now that concept has some statistical teeth. Extraordinary knowledge with respect to a given estimator is simply any knowledge not included in the Bayesian inference used to derive that estimator.

    It might at first seem like this is just an issue of semantics, but it's really a whole lot more. We're now no longer confined to talking about a single market line that has the property of "efficiency" to some specified degree, but rather we can talk about any number of market estimators each gauged by specific p and ν parameters. These estimators could represent current lines at any number of sportsbooks as well as any given bettor's own forecasts (the proviso being the conditional independence of the estimates or at least a gauge as to how the conditional joint probabilities of the estimates relate).

    Let's consider a hypothetical example of such a partition.

    I'll assume that readers are already familiar with the Binomial distribution, which gives the probability of occurrence of a specified number of "wins" over a specified number of bets each with a constant probability.

    One feature of the Binomial distribution as it relates to Bayesian inference is that of it having a conjugate prior, specifically the Beta distribution. A conjugate prior is really just a convenience when applying Bayesian inference. What it tells us in this case is that if we model our prior distribution of the Binomial probability parameter (which here is to say P(H) in the statement of Bayes' Theorem above) as a Beta distribution, then all our posterior distributions (the probability of the hypothesis given the evidence, P(H|E)) will also be Beta distributed (only reparameterized to take into account of the new information).

    This is especially convenient for the Binomial distributions because the uniform distribution is itself a special case of the Beta. This means that if absent any information we can fairly judge all win probabilities equal likely, then we can use additional information to continuously create new Beta distributions of win probabilities.

    Here's how it might work in practice (you can obtain specific values of the Beta distribution using the BETADIST() function in Excel).

    1. Consider a sporting event so new that bettors are completely ignorant of the distribution of win probabilities and so can do no better than assuming all win probabilities equally likely (i.e., prior win probability estimates are drawn from the uniform distribution over the interval(0, 1)). This by definition is equivalent to the Beta Distribution with parameters mean = 50% and our value parameter ν = 2 (as explained above).

      To make this compatible with Excel, there's a slightly different parameterization of the Beta that we'll be using.

      Specifically we'll consider parameters α and β (the usage of "β" being by convention and unrelated to the name of the Beta distribution itself).

      The mean and variance of the prior Beta distribution for probability parameter p are then given by:

      E(p) = α α + β
      Var(p) = α + β (α + β)^2 * (α + β + 1)

      The previously utilized ν parameter could now be reinterpreted as the sum of the parameters α + β yielding:
      E(p) = α ν
      Var(p) = E(p) * (1-E(p)) ν+1


      So by assuming a uniformly distributed prior (encapsulated by α = β = 1, with ν = α + β = 2) this gives us:

      E(p) = α α + β = 1 2 = 50%
      Var(p) = E(p)*(1-E(p)) ν + 1
      = 50%*50% 3
      = 1 12

      Which are just the mean and variance of the uniform distribution over support p ∈ (0%, 100%). The value parameter for this initial prior is ν = 2.
    2. Now let's say that after 100 games have been observed (all between different teams), the only information available is that out of those 100 games the home team won 55.

      Consistent with our beta conjugate prior to the binomial, we update our distribution as follows:

      αnew = αold + successes = 56
      βnew = βold + failures = 46
      νadded = 100
      νcumulative = 102

      This means that our new (posterior) distribution for home team win probability is distributed as:

      p ~ Beta(56, 46)

      This then implies:
      E(p) = α α + β = 56 56+46 ≈ 54.9020%
      The result for E(p) above is often referred to as the "rule of succession".

      Var(p) = 56*46 102*102 * (56 + 46 + 1)
      ≈ 0.2404%

      This would then imply a fair line of ~ -121.739 on the home team in the next match up.
    3. Now let's imagine that some additional information comes to light regarding the record of a specific team. Let's say that we discover that this team has won 17 of its last 20 games against fully typical opponents. If this team were next playing on the road what would the fair line on that game be?

      To answer this we'll need to perform another round of Bayesian estimation to yield a posterior distribution that takes into account this new information. Looking at it from the perspective of the home team:
      αnew = αold + successes = 56+3 = 59
      βnew = βold + failures = 46+17 = 63
      p ~ Beta(55+1+3, 45+1+17)
      νadded = 20
      νcumulative = 122

      E(p) = α/ν = 59/122 ≈ 48.3607%
      Var(p) = 59/122 * (1-59/122) /(123) ≈ 0.2030%
      This would then yield a fair line of +106.780 on the home team.
    4. Up to this point we've been solely considering historical categorical win frequencies as components to this framework. There's no reason that this need be the case in general, however. As long as a model may be used to estimate binary probabilities, then once we assign a "value" for those estimates in the form of our ν parameter (this can be obtained by considering the volatility of probability estimates) we can input these figures directly into our Bayesian inference model.

      This could be especially useful in inferring the value of a book's line based upon how they react to new information.

      So let's go back to the original 55/45 home away record and assume that new information has now come to light ... specifically regarding the composition of one of the teams.

      Once this information is released, Pinnacle now offers a no-vig line of even odds (and we'll assume for the sake of simplicity that they aren't biasing their market in any way).

      What can we infer from Pinnacle's actions? Well if we assume they're working off the same priors as we are, then we know that whatever the form of their new model, it managed to bring α and β up to equal footing. This will be useful in the next illustration.
    5. Let's say the next game was won by the home team (offered by Pinnacle at even odds) and the two teams were then slated to play another game following on the same home field.

      Further assume that this time around Pinnacle hangs a -100.6579 zero-vig line (p= 50.1639%) on the home team.

      What can we infer this go-around?

      Well this time there's actually quite a lot there.

      First we'll update our priors for the home/away model:

      αnew = αold + successes = 57
      βnew = βold + failures = 46
      νadded = 1
      νcumulative = 103

      E(p) = α/ν = 57/103 ≈ 55.3398%

      But now we're better poised to utilize Pinnacle's line signal.

      We already determined that Pinnacle's old values αpinnacle-old and β pinnacle-old values were equal. We can further state that Pinnacle's new αpinnacle and βpinnacle values must obey:

      αpinnacle / (αpinnacle + βpinnacle) = 50.1639%

      So by incorporating updates to the home away model (although admittedly without considering updates to the proprietary Pinnacle model) we also have:

      αpinnacle = αpinnacle-old + 1

      So putting it all together:

      αpinnacle / (αpinnacle + αpinnacle - 1) = 50.1639%
      αpinnacle = 50.1639%/(2 * 50.1639% - 1) = 153
      βpinnacle = 153 - 1 = 152
    6. This time, let's say we beat Pinnacle to market having just learned of a last moment game change. The upcoming game, it turns out, will for the first time in the history of the sport be played on neutral ground.

      Now obviously we’d want to jump on what had been the previous away line of +100.6579, but the trickier question here is what would be our estimate of the edge on the away team bet?

      In this case all we'd need to do be would subtract out the alpha and beta additions provided by the home/away model from the those inferred as Pinnacle's proprietary model parameters. This gives us:

      αnew = 153 - 56 = 97
      βnew = 152 - 45 = 107

      E(p) = 47.5490%
      This yields an edge on the away bet at +100.6579 of 5.2%.


    Now obviously the above relies on a very simplified set of assumptions, but it's presented to demonstrate the methodology of Bayesian inference. Even with a relatively paltry model, if one properly conditions one's prior forecasts on the information made implicitly available via publicly available lines, profit opportunity may abound.

    The important take away from all this (and hopefully this will the start of an ongoing discussion) is that new information needs to be continuously used to condition prior knowledge. To that end, it's imperative that when building a model one attempt to incorporate timely broad-based market information (e.g., widely available market indicative lines) into one's forecasts.

    Hopefully, I've provided a good starting point that yields food for additional thought.

  2. #2
    uva3021
    uva3021's Avatar Become A Pro!
    Join Date: 03-01-07
    Posts: 537
    Betpoints: 381

    very nice, can't wait to read and absorb the whole thing, learn so much from Ganchrow

    i noticed though you didn't clarify the meaning of the upside down A (for all values of some variable) for those not familiar with mathematical reasoning

    so p * 0 = 0 where p = probability is true for all (inverted A)values of p

  3. #3
    mathdotcom
    mathdotcom's Avatar Become A Pro!
    Join Date: 03-24-08
    Posts: 11,689
    Betpoints: 1943

    Those not familiar with mathematical reasoning won't be able to understand this anyways.

    There are a bunch of papers written on this; the simplest form is checking to see if various betting rules would turn a profit, and if so, is it statistically significant. Not surprisingly, blindly betting home underdogs and unders has not proven to be a very successful strategy.

    The more complicated form, yet extremely simplistic for a publishable paper, uses probit/logit models and simply runs the regression:

    Prob(Home Win) = Market Estimate of Probability of Home Win + whatever other available information you consider to be important (such as a Home Field dummy, the Total, the previous game's result, etc.) + e

    and checking to see if any of the other available information is statistically significant.

    Sadly, the above article would yield a better and more thorough paper than any of the academics have come up with. The famous author of Freakonomics, S. Levitt, uses data from an NFL handicapping tournament to try and make inferences about market efficiency - an idea which is of course ridiculous given tournaments lead to strategic situations that cause bettors to bet games they otherwise wouldn't.

    Ganch, if you want an article published in a (mediocre) journal you should e-mail some of the current economist dorks working on gambling markets: Andrew Weinbach, Rodney Paul, Woodland & Woodland, Raymond "Skip" Sauer, etc.

  4. #4
    Peregrine Stoop
    Peregrine Stoop's Avatar Become A Pro!
    Join Date: 10-23-09
    Posts: 869
    Betpoints: 779

    nice stuff

  5. #5
    JustinBieber
    JustinBieber's Avatar Become A Pro!
    Join Date: 05-16-10
    Posts: 324
    Betpoints: 149

    Asian books have a more efficient market than pinnacle? Could you list some of these books please I am intrigued.

  6. #6
    durito
    escarabajo negro
    durito's Avatar Become A Pro!
    Join Date: 07-03-06
    Posts: 13,173
    Betpoints: 438

    Quote Originally Posted by JustinBieber View Post
    Asian books have a more efficient market than pinnacle? Could you list some of these books please I am intrigued.
    SBObet for soccer handicaps, though pinny is trying to catch up.

  7. #7
    Tomato
    Tomato's Avatar Become A Pro!
    Join Date: 01-29-09
    Posts: 1,251
    Betpoints: 618

    "This report, by its very length, defends itself against the risk of being read."

    - Winston Churchill
    Nomination(s):
    This post was nominated 1 time . To view the nominated thread please click here. People who nominated: pico

  8. #8
    durito
    escarabajo negro
    durito's Avatar Become A Pro!
    Join Date: 07-03-06
    Posts: 13,173
    Betpoints: 438

    Yes, but those of that read it appreciate it more.

  9. #9
    pico
    USC ml
    pico's Avatar Become A Pro!
    Join Date: 04-05-07
    Posts: 27,321
    Betpoints: 1634

    Quote Originally Posted by Tomato View Post
    "This report, by its very length, defends itself against the risk of being read."

    - Winston Churchill

  10. #10
    byronbb
    byronbb's Avatar Become A Pro!
    Join Date: 11-13-08
    Posts: 3,067
    Betpoints: 2284

    9. information gained by watching ESPN

    Really? Or is this more about information the viewing public will interpret incorrectly?



  11. #11
    Ganchrow
    Nolite te bastardes carborundorum.
    Ganchrow's Avatar Become A Pro!
    Join Date: 08-28-05
    Posts: 5,011
    Betpoints: 1088

    Quote Originally Posted by byronbb View Post
    9. information gained by watching ESPN

    Really? Or is this more about information the viewing public will interpret incorrectly?
    I was just proposing a hypothetical to illustrate "weak-form" extraordinary knowledge.

  12. #12
    20Four7
    Timmy T = Failure
    20Four7's Avatar Become A Pro!
    Join Date: 04-08-07
    Posts: 6,703
    Betpoints: 4120

    Great to have you back Ganch...... I have to re read it in the morning.

  13. #13
    Data
    Data's Avatar Become A Pro!
    Join Date: 11-27-07
    Posts: 2,236

    I personally subscribe to the Adaptive Market Hypothesis. Your musings with the EMH, while sound, do not excite me the least bit because I think that the whole EMH framework leads nowhere and when it comes to the betting markets it is overwhelmingly more so.

    On a side note, it is somewhat ironic that such an avid Kelly scholar as yourself paid so much homage to Samuelson and Co.

    On a side side note, "beating the efficient market" is an oxymoron while I am in this for the money.

  14. #14
    TakeIt
    TakeIt's Avatar Become A Pro!
    Join Date: 04-23-10
    Posts: 776
    Betpoints: 25397

    Quote Originally Posted by Ganchrow View Post
    Let's start off by making a very general observation about betting markets: Betting markets are inherently competitive -- they reward winners and penalize losers. This creates a dynamic fundamentally different from what we see in, say, internet polling where there's no self-reinforcing drive to succeed.

    Outside the betting world and without money on the line there's little incentive for accuracy.

    Within the betting world, however, people make their opinions known with a signal more telling than their vote -- their money. Within the competitive marketplace it's democracy be damned and capitalism rules the roost.

    (Now certainly this signal can get corrupted by such forces as recreational bettors who'll blindly bet their home team at any odds or by games thrown by dishonest referees or players. Market economists account for these factors by assuming that the smart money (including both books and players betting for profit and with superior information flow and processing), will in totality eventually overwhelm the rest.)

    This leads to a framework known as (profit-wise) "market efficiency".

    Conceptually, the way we'll be considering what market efficiency as applied to sports betting says is that without "extraordinary knowledge" (including meta-knowledge) a player's expected profit from betting at the market no-vig line is on average zero.

    This definition still allows for many interpretations of market efficiency, running the continuum from so-called "strong form" (most restrictive in its assumptions as to market behavior) to "weak form" (least restrictive). But at its most fundamental the defining parameter comes down to little more than what exactly is meant by the characterization of knowledge as "extraordinary".

    Following is an indicative list of examples of types of knowledge going from those likely deemed extraordinary by all but adherents to the strongest forms of efficiency to those only so deemed (possibly) by proponents of the weakest:
    1. first-hand knowledge of game fixing
    2. insider information regarding a change in directive for officials
    3. insider information regarding a key player's health
    4. faster receipt and processing of new information
    5. faster response time following receipt of new information
    6. access to non-readily available (but still public) information
    7. novel quantitative processing of existing information
    8. any proprietary processing of existing information
    9. information gained by watching ESPN
    10. deeper fundamental understanding of motivations and factor facing one's home team or University

    Some strong form advocates would say that while they wouldn't in theory doubt the efficacy of some of the items on the above list, they would call in to question their general existence (or might just point to the infrequency with which such knowledge is found and so tend to consider them little more than just pathological cases).

    Personally, I'll just leave my own characterization as purposely vague with the eventual intention of providing some form of weak parameterization of the concept. I will say that given the above listing I'd probably be somewhere between the middle and the bottom.

    Another question one should ask would be what exactly represents a market line? Here there's a lot less debate. A market line should be one accessible (and already widely bet) to the betting public as a whole and one that won't be artificially held in place by a book limiting the number of bets by individual players (in other words after a reasonable period of time a market line should either move out of the way or accept another bet at the same price).

    The best examples for many sports would probably be Pinnacle, BetFair, and Matchbook. For certain geographically disparate sports, however, one would probably tend to find a better indication of the true market line with some of the Asian books. Generally speaking, very high limits, low juice, market prominence, and attitude towards professionals will jointly determine how well a given sports book's lines are indicative of the market as a whole.

    For the sake of argument we'll just say "the market" is represented by the inside line across geographically-appropriate "Pinnacle-like" books and exchanges. (Within what markets and under what circumstances Pinnacle itself would actually represent a "Pinnacle-like" book would be very much up for debate.)

    Getting right down to it, recall that we that declared the defining characteristic of market efficiency to be that a non-extraordinary player's expected profit from betting at the market no-vig line would on average be zero (or equivalently that on average E[player profit] = -theoretical hold, regardless of market side chosen).

    Recall that for a true win probability of p and a no-vig decimal line of d:
    E[profit] = p*d - 1 = 0
    d = 1 p
    But this definition of market odds is intuitively rather unsatisfying as it presupposes exact knowledge of win probability. Remember that the only defining characteristic of market efficiency to which we've limited ourselves is that (absent extraordinary information, which we discount for the time being) expected profit on bets placed at the zero-vig line will always be zero. But does this really require that the market knows the true win probability exactly all the time every time without margin for error?

    What we're trying to get at here are some determinations about how the market might set its lines (and although we keep referring to "the market" as a single entity, understand that even though the market line might be reflected at only a single book, it the market as a whole represents a conglomeration of opinions across professional players and books. The mathematical mechanics of how "the market" might handle such a conglomeration this we won’t delve into.

    I expect that anyone who's already gotten this far is already familiar with the concept of expectation as it relates to, say, value. Specifically, the expected value of a bet could be described as the theoretical average profit per game were the game to be repeated an infinite number of times (at least that's how a "frequentist" would describe it).

    So we'll write:
    E[p̂] = 1 d
    d = 1 E[p̂]
    Where the caret-looking symbol lurking over (or nearly over) the p is known by the highly complex term "hat". In probability, a hat-ed variable almost always represents an estimate of a distribution parameter.

    But how exactly should we interpret the seemingly innocuous term "E[p̂]"? Well all we're doing is applying the concept of expectations to a random variable.

    It's just that in this case that random variable happens not to be an estimate of value but rather of probability ... more specifically the market's estimate of the true win probability of the game.

    Think about it in a manner similar to expected value. If we repeated the game an infinite number of times and then repeated that set of infinite repetitions an infinite number of times what we'd see is that across each one of those infinite sized game-groups, the frequency of wins was on average the same as the expected win probability (note that this abstraction to an infinite number of infinitely sized game groupings isn't necessary and is only presented with the hope that it might aid in visualization).

    This is an extremely important concept. We're not forcing the assumption that the market is always correct in its probability estimates ... just that its probability estimates are correct on average.

    And there's another way to consolidate the above:
    E[p̂] = p
    So in other words, the expected value of the market estimate of probability equals the true probability. In more technical terms this statement can be read to say that p̂ is an unbiased estimator of p. So while a given probability estimate might be wrong in any single case it's still correct on average. This is perfectly analogous to expected value.

    (FWIW, note that this is not an equivalent statement to the similar E[1/p̂] = 1/p, so in other words while we’ve concluded based on our assumptions that the zero-vig implied probability is an unbiased estimator of true probability, we can't in general conclude that the zero-vig line will be an unbiased estimator of the true line.)

    So now that we've looked at the expected value of the market probability estimate, the next logical meta-statistic would be the variance of the market probability estimate. Once again we're not deviating from well-trod territory -- this is analogous to not only talking about the expected value of a bet, but also talking about the variance of the value of a bet.

    For a reason that should soon become I'm going to switch over to time subscripts and propose a very simple model of market efficiency such that the market estimate of win probability at ordinal time t, p̂t (where t = 0, represents the present and t = Τ represents the market close) obeys:
    1. E(p̂t) = p ∀ 0 ≤ t ≤ Τ
    2. Var(p̂s) > Var(p̂t) > 0 ∀ s > t ≥ Τ

    So what this is saying is that at any moment in time, the reciprocal of the market line (which is to say, p̂t) is an unbiased estimator of true win probability and that the variance of that estimate is strictly decreasing as game time approaches.

    As a corollary to 1) above we can write:
    E(p̂t) = E(p̂0) ∀ 0 ≤ t
    Or in other words the expectation of the implied market probability at any point in the future is just the current implied market probability. This follows directly from the unbiasedness of the probability estimator across time.

    Let's step back and consider this all for a moment. We're saying that at any point in time between now and game start the market estimate of probability will on average be correct and will become continuously more accurate (meaning less variance) as time progresses.

    A good way to intuit this would be in terms of weather forecasting, say the probability of precipitation on a given day. In theory we might say that every forecast we hear we believe to be unbiased (correct on average) and that we further expect the reliability of each forecast to increase (lower variance) as we get closer to the day for which we're forecasting.

    Moving back to statement 1) above, let's try to find some sort of additional functional bound on the market estimator. One particularly apparent bound should be fairly easy to spot. If we're estimating some probability p, and know that that estimate will be correct on average what’s the noisiest possible probability estimate that will still be unbiased?

    It should eventually become apparent that the maximum variance estimator p̂, would simply be the estimator that takes on the value p̂=100% with probability=p, and the value of p̂=0% with probability=(1-p).

    This is just another way of saying that p̂ were Bernoulli distributed with parameter p. (Once again think of weather forecasting ... a Bernoulli forecast of this type would always offer either a 100% or 0% probability of rain on a given future day, and while it might be correct on average -- and remember that the true probability of rain on a future is never really 0% or 100% -- it'll certainly be less valuable in picking a future day for a picnic than a lower variance probability estimate).

    We know that a Bernoulli-distributed variable with parameter probability parameter p has variance = p*(1-p), so when taken in conjunction with condition 2) which says that future variance of a given estimate will always be lower than the current variance, this gives another bound on Var(p̂2). Specifically, 2) implies that maximum variance of p̂t could only be achieved at t=&-infin;, which means that for all finite t:
    Var(p̂t) < p*(1-p)
    Recalling that for any random variable X:
    Var(X) = E[X2] - E[X]2
    We could also say:
    E[p̂2] < p*(1-p) + p2 ≡ p
    These are by necessity rather weak statements, but they nevertheless do give lend a bit more probabilistic rigor to the notion of market efficiency by allowing us to place functional bounds on the estimator p̂:
    E(p̂) = p
    Var(p̂) < p*(1-p)
    But this upper bound on Var(p̂) is the result of assuming a maximum variance estimator of p, which was that obtained by the Bernoulli distribution and a high variance estimator like just doesn't pass the smell test. A probability of 0%/100% would yield a bet at +∞ on one team or the other. Whatever assumptions we might accept that's just a full out absurd result. So we'd really hope that we can improve upon that.

    But where to turn? Given that the Beta distribution is the conjugate prior of the binomial it certainly might be convenient to look there.

    Recalling our vast knowledge of probability distributions (or Wikipedia) we can say that if p̂ is Beta distributed with mean p and Beta shape parameter ν, then the variance of p̂ is given by:

    1. Var(p̂) = p*(1-p) 1+ν


    It's via this parameter ν (the Greek letter nu) that we can start thinking of providing a parameterization of the concept of market efficiency.

    Specifically ν → 0, represents the maximum variance unbiased estimator of p, while ν → ∞ would yield constant win probability p̂ ≡ p. And anywhere else in between is fair game.

    This is actually a meaningful result. What we've done is use a parameter from a well-known distribution to place an equality bound on the variance of the market probability estimate effectively parameterizing the degree of efficiency for a given market. Of course we're still assuming a magically unbiased line that's even more magically taking into account all market knowledge ... but patience is a virtue.

    To continue further, we’ll consider our model from the perspective of Bayesian inference.

    What is Bayesian inference you might ask?

    Well it's really a different way of looking at probability that allows for a different methodology in forecasting. Recall that I earlier wrote that a frequentist would view the expected value of a bet as the average profit per game were the bet to be repeated an infinite number of times. Well that's just not how a Bayesian sees the world.

    A Bayesian considers the probability of an outcome as a (possibly subjective) measure of the degree of informed belief in that outcome. We talk about "informed" belief because the process of Bayesian inference involves updating one's prior beliefs based on the availability of new evidence.

    A Bayesian doesn't in general think about hypothetical frequencies of an event given a hypothetical infinite number of repetitions because in general events can't be repeated an infinite number. To estimate outcome probability what a Bayesian does is gauge prior knowledge of that event and then update that knowledge as future information becomes available.

    Note that this methodology doesn't hold value for all types of experiments as for some events can know everything there is to know about it a priori. These events (take the game of Craps for instance) are frequentist in nature meaning that there is no value to new information (although Craps could deemed be otherwise were we to suspect either an unfair game or the presence of a skilled dice roller).

    This can be a particular convenient tool when looking at the progression of a betting line. One can build a forecast in any way one chooses and then continually reevaluate that forecast based on upon the availability of new evidence (e.g., a change in the market line).

    By way of contrast, a non-Bayesian might place a bet at a line of +3, and then after observing the line move to +5 simply declare his original bet "bad" in that the market hadn't backed up his opinion. A Bayesian, on the other hand would realize that his bet was made conditioned only on the information he had available at the time the bet was made (which would have only included the then current line), and while he would almost certainly view the bet as "unfortunate", he could accept that the bet was still a good bet at the time it was placed. Certainly he'd use the new information to revise his current opinions on game probabilities, and going a step further might even use it to discount the value of his model, but a Bayesian can accept that a decision might be perfectly valid at the time it was made, even as new information sheds doubt on it in hindsight.

    At the heart of all Bayesian inference is what's known as Bayes' theorem. This is stated as follows:
    Let H = A hypothesis (for example the hypothesis that win probability = a specific value p̂)
    Let E = The evidence

    Prob(H|E) = P(E|H) * P(H) / P(E)
    What this says is that the probability of a hypothesis (which we can extend to a probability distribution) given some set of evidence is proportional to how likely we'd be to view that evidence assuming our hypothesis true times the probability of the hypothesis based solely on prior information (i.e., without benefit of any new evidence).

    Armed with this added understanding, we can meaningfully back off from our prior view of market efficiency by relaxing the global assumption of p̂ being a strictly unbiased estimator of p and replace it with the weaker assumption:
    p̂(K) = E(p | Κ)
    where Κ represents a given "knowledge space" comprised not only of timely and historical event and market data but also of specific prior knowledge of relevant deterministic functional relationships between that data and p̂, the variable of interest.

    Note a couple of things about this equation.

    Firstly, we've moved our p̂ estimator outside of the expectation and moved the parameter p representing "true" market probability inside the expectation. This is consistent with a Bayesian framework. The actual probability value is itself viewed as a random variable, while the estimator of p is viewed as a dependant variable. This is because Bayesian true probability isn't viewed as a discrete measurable quantity and can only viewed in terms of information upon which estimates are conditioned.

    Secondly, note that we're now making explicit the functional relationship between our estimator of p̂ and the knowledge space. This reflects the notion that the estimator is only as valid as the knowledge upon which it's based. Also included in K would ν(K), which once again represents a measure of inherent "value" of the knowledge space, only in this case relating that value not only to the variance of the estimate but also to its mean.

    This is virtually akin to stating that p̂ is a minimum mean square error (MMSE) estimator of p conditioned on "prior knowledge". It isn't exactly the same because we're necessitating explicit inclusion of "relevant deterministic functional relationships" within the space. A true MMSE estimator would need to already take into account all possible functional relationships between information pieces.

    Still what this allows for is an explicit decomposition of "market probability" into any number of discrete conditionally independent knowledge packets. What’s the relevance of this? Well it’s actually quite substantial. Remember our vague "extraordinary knowledge" constraint? Well now that concept has some statistical teeth. Extraordinary knowledge with respect to a given estimator is simply any knowledge not included in the Bayesian inference used to derive that estimator.

    It might at first seem like this is just an issue of semantics, but it's really a whole lot more. We're now no longer confined to talking about a single market line that has the property of "efficiency" to some specified degree, but rather we can talk about any number of market estimators each gauged by specific p and ν parameters. These estimators could represent current lines at any number of sportsbooks as well as any given bettor's own forecasts (the proviso being the conditional independence of the estimates or at least a gauge as to how the conditional joint probabilities of the estimates relate).

    Let's consider a hypothetical example of such a partition.

    I'll assume that readers are already familiar with the Binomial distribution, which gives the probability of occurrence of a specified number of "wins" over a specified number of bets each with a constant probability.

    One feature of the Binomial distribution as it relates to Bayesian inference is that of it having a conjugate prior, specifically the Beta distribution. A conjugate prior is really just a convenience when applying Bayesian inference. What it tells us in this case is that if we model our prior distribution of the Binomial probability parameter (which here is to say P(H) in the statement of Bayes' Theorem above) as a Beta distribution, then all our posterior distributions (the probability of the hypothesis given the evidence, P(H|E)) will also be Beta distributed (only reparameterized to take into account of the new information).

    This is especially convenient for the Binomial distributions because the uniform distribution is itself a special case of the Beta. This means that if absent any information we can fairly judge all win probabilities equal likely, then we can use additional information to continuously create new Beta distributions of win probabilities.

    Here's how it might work in practice (you can obtain specific values of the Beta distribution using the BETADIST() function in Excel).

    1. Consider a sporting event so new that bettors are completely ignorant of the distribution of win probabilities and so can do no better than assuming all win probabilities equally likely (i.e., prior win probability estimates are drawn from the uniform distribution over the interval(0, 1)). This by definition is equivalent to the Beta Distribution with parameters mean = 50% and our value parameter ν = 2 (as explained above).

      To make this compatible with Excel, there's a slightly different parameterization of the Beta that we'll be using.

      Specifically we'll consider parameters α and β (the usage of "β" being by convention and unrelated to the name of the Beta distribution itself).

      The mean and variance of the prior Beta distribution for probability parameter p are then given by:
      E(p) = α α + β
      Var(p) = α + β (α + β)^2 * (α + β + 1)
      The previously utilized ν parameter could now be reinterpreted as the sum of the parameters α + β yielding:
      E(p) = α ν
      Var(p) = E(p) * (1-E(p)) ν+1


      So by assuming a uniformly distributed prior (encapsulated by α = β = 1, with ν = α + β = 2) this gives us:
      E(p) = α α + β = 1 2 = 50%
      Var(p) = E(p)*(1-E(p)) ν + 1
      = 50%*50% 3
      = 1 12
      Which are just the mean and variance of the uniform distribution over support p ∈ (0%, 100%). The value parameter for this initial prior is ν = 2.
    2. Now let's say that after 100 games have been observed (all between different teams), the only information available is that out of those 100 games the home team won 55.

      Consistent with our beta conjugate prior to the binomial, we update our distribution as follows:
      αnew = αold + successes = 56
      βnew = βold + failures = 46
      νadded = 100
      νcumulative = 102

      This means that our new (posterior) distribution for home team win probability is distributed as:
      p ~ Beta(56, 46)
      This then implies:
      E(p) = α α + β = 56 56+46 ≈ 54.9020%
      The result for E(p) above is often referred to as the "rule of succession".

      Var(p) = 56*46 102*102 * (56 + 46 + 1)
      ≈ 0.2404%
      This would then imply a fair line of ~ -121.739 on the home team in the next match up.
    3. Now let's imagine that some additional information comes to light regarding the record of a specific team. Let's say that we discover that this team has won 17 of its last 20 games against fully typical opponents. If this team were next playing on the road what would the fair line on that game be?

      To answer this we'll need to perform another round of Bayesian estimation to yield a posterior distribution that takes into account this new information. Looking at it from the perspective of the home team:
      αnew = αold + successes = 56+3 = 59
      βnew = βold + failures = 46+17 = 63
      p ~ Beta(55+1+3, 45+1+17)
      νadded = 20
      νcumulative = 122

      E(p) = α/ν = 59/122 ≈ 48.3607%
      Var(p) = 59/122 * (1-59/122) /(123) ≈ 0.2030%
      This would then yield a fair line of +106.780 on the home team.
    4. Up to this point we've been solely considering historical categorical win frequencies as components to this framework. There's no reason that this need be the case in general, however. As long as a model may be used to estimate binary probabilities, then once we assign a "value" for those estimates in the form of our ν parameter (this can be obtained by considering the volatility of probability estimates) we can input these figures directly into our Bayesian inference model.

      This could be especially useful in inferring the value of a book's line based upon how they react to new information.

      So let's go back to the original 55/45 home away record and assume that new information has now come to light ... specifically regarding the composition of one of the teams.

      Once this information is released, Pinnacle now offers a no-vig line of even odds (and we'll assume for the sake of simplicity that they aren't biasing their market in any way).

      What can we infer from Pinnacle's actions? Well if we assume they're working off the same priors as we are, then we know that whatever the form of their new model, it managed to bring α and β up to equal footing. This will be useful in the next illustration.
    5. Let's say the next game was won by the home team (offered by Pinnacle at even odds) and the two teams were then slated to play another game following on the same home field.

      Further assume that this time around Pinnacle hangs a -100.6579 zero-vig line (p= 50.1639%) on the home team.

      What can we infer this go-around?

      Well this time there's actually quite a lot there.

      First we'll update our priors for the home/away model:
      αnew = αold + successes = 57
      βnew = βold + failures = 46
      νadded = 1
      νcumulative = 103

      E(p) = α/ν = 57/103 ≈ 55.3398%
      But now we're better poised to utilize Pinnacle's line signal.

      We already determined that Pinnacle's old values αpinnacle-old and β pinnacle-old values were equal. We can further state that Pinnacle's new αpinnacle and βpinnacle values must obey:
      αpinnacle / (αpinnacle + βpinnacle) = 50.1639%
      So by incorporating updates to the home away model (although admittedly without considering updates to the proprietary Pinnacle model) we also have:
      αpinnacle = αpinnacle-old + 1
      So putting it all together:
      αpinnacle / (αpinnacle + αpinnacle - 1) = 50.1639%
      αpinnacle = 50.1639%/(2 * 50.1639% - 1) = 153
      βpinnacle = 153 - 1 = 152
    6. This time, let's say we beat Pinnacle to market having just learned of a last moment game change. The upcoming game, it turns out, will for the first time in the history of the sport be played on neutral ground.

      Now obviously we’d want to jump on what had been the previous away line of +100.6579, but the trickier question here is what would be our estimate of the edge on the away team bet?

      In this case all we'd need to do be would subtract out the alpha and beta additions provided by the home/away model from the those inferred as Pinnacle's proprietary model parameters. This gives us:
      αnew = 153 - 56 = 97
      βnew = 152 - 45 = 107

      E(p) = 47.5490%
      This yields an edge on the away bet at +100.6579 of 5.2%.


    Now obviously the above relies on a very simplified set of assumptions, but it's presented to demonstrate the methodology of Bayesian inference. Even with a relatively paltry model, if one properly conditions one's prior forecasts on the information made implicitly available via publicly available lines, profit opportunity may abound.

    The important take away from all this (and hopefully this will the start of an ongoing discussion) is that new information needs to be continuously used to condition prior knowledge. To that end, it's imperative that when building a model one attempt to incorporate timely broad-based market information (e.g., widely available market indicative lines) into one's forecasts.

    Hopefully, I've provided a good starting point that yields food for additional thought.
    is all of this just another way of saying, "no edge no bet?"
    Nomination(s):
    This post was nominated 1 time . To view the nominated thread please click here. People who nominated: 135steward

  15. #15
    Ganchrow
    Nolite te bastardes carborundorum.
    Ganchrow's Avatar Become A Pro!
    Join Date: 08-28-05
    Posts: 5,011
    Betpoints: 1088

    Quote Originally Posted by Data View Post
    I personally subscribe to the Adaptive Market Hypothesis. Your musings with the EMH, while sound, do not excite me the least bit because I think that the whole EMH framework leads nowhere and when it comes to the betting markets it is overwhelmingly more so.

    On a side note, it is somewhat ironic that such an avid Kelly scholar as yourself paid so much homage to Samuelson and Co.
    weak
    On a side side note, "beating the efficient market" is an oxymoron while I am in this for the money.
    Oh Data, Data. I'm surprised at you.

    That's not it at all.

    I think that the conclusions drawn above should (if believed) be heartening to a modeler such as yourself.

    We're making explicit the notion that one doesn't need to be smarter than the market at each and every turn in order to be profitable. But rather by simply establishing a "knowledge niche" as a prior model and then updating it via market info, opportunities can abound.

    A simple hypothetical:

    Let's say you find such a knowledge niche at a particular categorical level of NCAA Football that conditionally models win probability as some function of field goal this or that blah blah blah.

    Now obviously by its nature this is not going to be an all-encompassing model, but that doesn't imply it need be valueless.

    Let's say that (in the parlance of the above write up) you assign a value parameter of ν[sub]model[/hcc] = 100 to this model (which you've determined based on prediction volaitility) and have also backed out a cumulative ν[sub]pinnacle open[/hcc] = 2,400 for Pinnacle openers.

    Pinnacle hangs the game at PK Even, while your model predicts a 60% win probability.

    After conditioning you find:

    E(p) = 1,200+60 2,400+100 = 50.40%.
    Stdev(p) = 4.875%

    prior p-value of 60% win probability forecast
    =BETADIST(50.40%, 61, 41) ≈ 2.8%
    Now this is obviously a tiny edge (EV = 0.8%, probably within the model's margin of error) that corresponds to a forecast frequency of fewer than 28 games per thousand ... but that's very much beside the point I think.

    I could have juked the numbers on the hypothetical and presented some seemingly wondrously promising result, but that isn't where I'm trying to take this.

    Using Bayesian inference in conjunction with certain assumptions regarding weak-form market efficiency, a quantitative bettor doesn't need to be thinking only in terms of big, manly, all-encompassing models. Rather a savvy modeler can look for knowledge niches and create small, easily parametrizable and estimable models that take advantage of gaps overlooked by the market.

    In the above example (and returning to frequentism for the purpose of illustration) Pinnacle's opener implicitly had a prior 95% confidence interval of 50%±2%, while the hypothetical FG model would have yielded a 95% confidence interval about 50% of ±9.6%. Now do I think that I could reasonably come up with a CFB FG model that was able to determine win probabilities within an (α = 5%) accuracy of ±9.6ish% for bets near even?

    Well actually now that I think about it probably not. But that's just a stupid example anyway.

    Think about all the models you've created over the years -- ones that you recognized had some predictive power but just couldn't seem to beat the number. Maybe they came close and hovered around break-even or maybe just drained slowly.

    Some of you know what I'm talking about. (I know Data does.)

    Well why not dust off some of those old babies and take another look from a Bayesian ineference standpoint?

    Or if you've never built a model but have been wanting to try why not just start small? Stick to an area with which you're already familiar and where you feel you'll be able to bring novel prior knowledge to market.

    You're almost certainly never going to build a tool that will reliably beat Pinnacle NFL closers. So stop trying (at least for now) and realize that there are other ways to be profitable in quantitative modeling.

    You say you already have a broad-based NFL model? Well why not try conditionally orthogonalizing it with Pinnacle lines and then reintroducing those lines as Bayesian updates?

    Trust the market outside the borders of your knowledge in order to better exploit it from within.

  16. #16
    thebayareabeast
    thebayareabeast's Avatar Become A Pro!
    Join Date: 04-22-10
    Posts: 1,475

    I bookmarmed this to come back and read it later

    Txs

  17. #17
    Data
    Data's Avatar Become A Pro!
    Join Date: 11-27-07
    Posts: 2,236

    I am sorry to disappoint, I am not as smart as you are. You must be seeing something what I do not.

    I think I did look into something similar if not the same and I found that if I trusted the line too much I would have to take many positions in derivatives that my models consider -EV and I did not like that at all. I am an arrogant SOB after all.
    Last edited by Data; 07-20-10 at 02:33 AM.

  18. #18
    MonkeyF0cker
    Update your status
    MonkeyF0cker's Avatar Become A Pro!
    Join Date: 06-12-07
    Posts: 12,144
    Betpoints: 1127

    Quote Originally Posted by Data View Post
    I personally subscribe to the Adaptive Market Hypothesis. Your musings with the EMH, while sound, do not excite me the least bit because I think that the whole EMH framework leads nowhere and when it comes to the betting markets it is overwhelmingly more so.

    On a side note, it is somewhat ironic that such an avid Kelly scholar as yourself paid so much homage to Samuelson and Co.

    On a side side note, "beating the efficient market" is an oxymoron while I am in this for the money.
    One could certainly employ certain aspects of the Noisy Market Hypothesis to sports markets as well. I have to agree that the AMH is more suited to these markets than the EMH. However, this is certainly a good read and the Bayesian inference examples that were set forth have valuable applications in sports betting.

  19. #19
    MonkeyF0cker
    Update your status
    MonkeyF0cker's Avatar Become A Pro!
    Join Date: 06-12-07
    Posts: 12,144
    Betpoints: 1127

    LOL. I guess I should have refreshed the page before responding.

  20. #20
    uva3021
    uva3021's Avatar Become A Pro!
    Join Date: 03-01-07
    Posts: 537
    Betpoints: 381

    Most EV bettors would have to apply various measures of Bayesian philosophy to some degree lest one alters his entire approach to handicapping based on one single outcome

    i know this is merely semantics, but the word unbiased is rather strong when dealing with markets, because pinnacle or some other oddsmakers are yet constrained to their modeling formulated by their own observations and experiences

    so sportsbooks may be unbiased in their usage of the information to convene a line, but the information modeled arrived from an observer's aggregation of such information

    maybe objective would be more appropriate, which might suggests a weaker connection between conscious and unconscious market manipulation, though allows for line appropriation to be tethered to the linesmaker's relative perception of information

    you are infinitely smarter than I am Ganchrow, I only bring it up because it is very similar to a book I'm currently reading, how scientific knowledge and logic is shaped by the delicate process of carefully selecting terminology, and objectivity vs unbiased is one of the topics

  21. #21
    Wrecktangle
    Wrecktangle's Avatar Become A Pro!
    Join Date: 03-01-09
    Posts: 1,524
    Betpoints: 3209

    Nice post, like others I'll come back and reread when more time.

    Ganch makes one post (OK, it's a large post) on Bayesian Inference and gets all sorts of crowd noise, whereas I've been yapping about it for two years now with no notice. I guess I've got to learn how to insert the symbolic math into my posts to have any effect.

  22. #22
    runnershane14
    runnershane14's Avatar Become A Pro!
    Join Date: 07-23-07
    Posts: 803
    Betpoints: 84

    ive noticed wreck

  23. #23
    uva3021
    uva3021's Avatar Become A Pro!
    Join Date: 03-01-07
    Posts: 537
    Betpoints: 381

    Quote Originally Posted by Wrecktangle View Post
    Nice post, like others I'll come back and reread when more time.

    Ganch makes one post (OK, it's a large post) on Bayesian Inference and gets all sorts of crowd noise, whereas I've been yapping about it for two years now with no notice. I guess I've got to learn how to insert the symbolic math into my posts to have any effect.
    http://www.sitmo.com/latex/

  24. #24
    MonkeyF0cker
    Update your status
    MonkeyF0cker's Avatar Become A Pro!
    Join Date: 06-12-07
    Posts: 12,144
    Betpoints: 1127

    Quote Originally Posted by uva3021 View Post
    Most EV bettors would have to apply various measures of Bayesian philosophy to some degree lest one alters his entire approach to handicapping based on one single outcome

    i know this is merely semantics, but the word unbiased is rather strong when dealing with markets, because pinnacle or some other oddsmakers are yet constrained to their modeling formulated by their own observations and experiences

    so sportsbooks may be unbiased in their usage of the information to convene a line, but the information modeled arrived from an observer's aggregation of such information

    maybe objective would be more appropriate, which might suggests a weaker connection between conscious and unconscious market manipulation, though allows for line appropriation to be tethered to the linesmaker's relative perception of information

    you are infinitely smarter than I am Ganchrow, I only bring it up because it is very similar to a book I'm currently reading, how scientific knowledge and logic is shaped by the delicate process of carefully selecting terminology, and objectivity vs unbiased is one of the topics
    You're talking about a completely different form of bias. What he's talking about is market bias. Assuming that a market is efficient, the true price (or probability in the case of sports markets) is not always equal to that of the current market price. However, deviations from fair value are simply a function of random walk and no investor could consistently find value in the market price.
    Last edited by MonkeyF0cker; 07-20-10 at 10:21 AM.

  25. #25
    MonkeyF0cker
    Update your status
    MonkeyF0cker's Avatar Become A Pro!
    Join Date: 06-12-07
    Posts: 12,144
    Betpoints: 1127

    I should mention that it's a product of the Efficient Market Hypothesis that Data alluded to earlier.

  26. #26
    mathdotcom
    mathdotcom's Avatar Become A Pro!
    Join Date: 03-24-08
    Posts: 11,689
    Betpoints: 1943

    Quote Originally Posted by Wrecktangle View Post
    Nice post, like others I'll come back and reread when more time.

    Ganch makes one post (OK, it's a large post) on Bayesian Inference and gets all sorts of crowd noise, whereas I've been yapping about it for two years now with no notice. I guess I've got to learn how to insert the symbolic math into my posts to have any effect.
    Most of your posts are way off topic and apply theory in completely incomprehensible ways (eg. game theory mortgage discussions??)

  27. #27
    thebayareabeast
    thebayareabeast's Avatar Become A Pro!
    Join Date: 04-22-10
    Posts: 1,475

    Market efficiency and bestiality probability

  28. #28
    That Foreign Guy
    I got sunshine in a bag
    That Foreign Guy's Avatar Become A Pro!
    Join Date: 07-18-10
    Posts: 432
    Betpoints: 3069

    My head asplode.

    Actually, it looks like really good stuff, I'm just too tired to give it the attention it deserves so this is a bookmarky post.

  29. #29
    Wrecktangle
    Wrecktangle's Avatar Become A Pro!
    Join Date: 03-01-09
    Posts: 1,524
    Betpoints: 3209

    Quote Originally Posted by mathdotcom View Post
    Most of your posts are way off topic and apply theory in completely incomprehensible ways (eg. game theory mortgage discussions??)
    I described how one can use GT in sports and described the best GT book to explain the theory on the market. You need to change your name.
    Last edited by Justin7; 07-21-10 at 11:30 AM.

  30. #30
    mathdotcom
    mathdotcom's Avatar Become A Pro!
    Join Date: 03-24-08
    Posts: 11,689
    Betpoints: 1943

    Quote Originally Posted by Wrecktangle View Post
    I described how one can use GT in sports and described the best GT book to explain the theory on the market. You need to change your name to Can't read or comprehend dot com.
    You recommended a game theory textbook to explain how a market works?

    The correct answer is to pick up an Economics 100 textbook and read about demand and supply.

    What does game theory have to do with finding +EV bets? The only relation I can think of is Justin7's post about weak openers and there being only a handful of people who know they're weak (= a prisoner's dilemma as all are better off waiting for the limits to increase before popping the weak number but, expecting others to pop it first, none can resist jumping on it quickly).

  31. #31
    Wrecktangle
    Wrecktangle's Avatar Become A Pro!
    Join Date: 03-01-09
    Posts: 1,524
    Betpoints: 3209

    This is my last post to you as the mods are now sending ME notes on trash talk. What I meant to say is: that particular GT book is the best GT book on the market. Research my posts on GT and you'll see where I give an instance of where GT is used in sports.

    Have a nice "9 to 5" life. I'm done wasting time with you.

  32. #32
    Data
    Data's Avatar Become A Pro!
    Join Date: 11-27-07
    Posts: 2,236

    Quote Originally Posted by Ganchrow View Post
    E(p) = 1,200+60 2,400+100 = 50.40%.
    Stdev(p) = 4.875%
    How did you you get 4.875%?

  33. #33
    Ganchrow
    Nolite te bastardes carborundorum.
    Ganchrow's Avatar Become A Pro!
    Join Date: 08-28-05
    Posts: 5,011
    Betpoints: 1088

    Quote Originally Posted by Data View Post
    How did you you get 4.875%?
    I should have explicitly noted that that the standard deviation of p was intended a prior for the hypothesis of a 60% win probability and without regard to the posterior Pinnacle line evidence.

    Stdev(p) = 60% * (1-60%) / (100 + 1) ≈ 4.875%
    Poor presentation on my part.

  34. #34
    Pokerjoe
    Pokerjoe's Avatar Become A Pro!
    Join Date: 04-17-09
    Posts: 704
    Betpoints: 307

    Quote Originally Posted by TakeIt View Post
    is all of this just another way of saying, "no edge no bet?"
    No.

  35. #35
    splash
    splash's Avatar Become A Pro!
    Join Date: 05-25-09
    Posts: 38

    Does anyone have a recommendation for books on Bayesian inference or even just statistical data analysis in general? I would like to read up more about this.

12 Last
Top