Hi,
I've been spending some time on compiling a database over past results (US and European tours since 1995), Betfair historical data (all tournaments/markets the last two years to start with = ~12000 bets/month on average) as well as retrieving current odds via Betfair free API. Now I need to start working on actually doing something with all this data
So I have two requests:
1)
I'm looking for help and/or resources on how to creating a model for, I guess, normalizing different types of data so it can be compared player vs player. The result would be some sort of power rating.
(something like the Sagarin index http://www.golfweekrankings.com/temp...lt.asp?t=world which I guess I would rip if it had been published in its total, which AFAIK it hasn't)
I have this (partial) example mockup data:
Player 1
Avg scoring 69.3
Putting GIR 1.73
Top 10s 2
Player 2
Avg scoring 68.5
Putting GIR 1.81
Top 10s 4
Player 3
Avg scoring 71.5
Putting GIR 1.89
Top 10s 0
I've made a rough outline in Excel of a starting point on the weight of each data type (adjusted average scoring, driving total, course form etc iterated over categories all/last30/last10/last3/last). But how is the model/math from going from above data to:
Player 1
Rating 91/100
Player 2
Rating 93/100
Player 2
Rating 86/100
2)
Given that you have this rating, the next step is to translate this into probabilities. So how do I get from 91 vs. 93 rating to 65%/35% win probabilities in a head to head situation. (or in another case, 50 players ratings into outright win probabilities)
With this I intend to create a system for running the model many times, with different weight parameters, against historical odds and results to see if I could've found value in the past. If so I should be able to find +EV spots in the future (I hope) If not, at least I learned a ton of Office, C#, SOAP, SQL Server and LINQ programming
Any help on these matters would be appreciated.
//pompano