I did a little analysis of FBG's projections and thought I'd share the results with whomever was interested. I've uploaded a pdf to mediafire that I hope is accessible (never used it before). First, to explain the figure, I took Dodds season total projections and divided by his projected number of games. I then plotted that vs the players actual points per game (as determined via the data dominator). I did this mostly because just looking at season totals would have a great deal of complications due to injuries and outright replacements. In addition most fantasy leagues are set up to make weekly choices for a lineup so I'd expect per game performance to be more useful to us. The R^2 values that are listed next to the graphs can be interpreted as the percent of variation in Actual performance that is explained by the projections.
Perhaps more clearly, essentially 60% of the variability in players actual points per game can be explained by footballguys pre-season projections! I have done no comparison across other websites but that is extremely impressive to me as someone who has examined a great deal of data types.
http://www.mediafire.com/?c9tgjnplgmklxj1 - All player projections vs actual results possible using FBG Data dominator (QB=32 players, RB=123 players, WR= 136 players, TE= 81 players
http://www.mediafire.com/?142p1z4lwqnqc2x - Same format of graph looking only at top starters (QB=12, RB=24, WR=36,TE=12).
However, The in season projections are perhaps less promising in usefulness. The following are a pooled data set of all projection-actual pairs for wr/rb/te (note no qb's) using dodds weekly projections. To read the following graph projection value of 5 means all projections between 5.00 and 5.99 etc. The number of projections is the number of times Dodds projected a player to score in that range on the season. It's a graph of median values (50% of actual below and 50% above), and the bars around each point are the Inner quartile range (25 percentile to 75th percentile). In essence 50% of all actual scores were between those ranges (Yes the ranges are that big!).
http://www.mediafire.com/?dwptr8t0o8ldnas
The means (not displayed on graph) of the lower projection values ( about 1 thru 10) are quite accurate as you can see by the above table. Which is to say if Dodds predicts a player will score in the 6.00-6.99 range the mean value of actual responses was 6.37 for example. In that projection range the actual values not only keep increasing, but they are accurate to the projected values themselves. Once you reach projections higher than 10 however, the data becomes more erratic. This is possible due to the much smaller sample sizes involved but nonetheless limits the usefulness of discerning between players projected at more than 10 points in a given week.
Another thing to be considered is that this data is a pooled grouping of 3 different positions (wr te and rb) the relationships above may be more or less accurate depending on position.
I think the following graph does a good job of showing the value of the projections (for RB's at least, more to follow). The legend might be a little confusing. Basically its 3 lines that show the following information: The green line is the season average of PPG at a given rank. So the 1st rank is the average of all of the highest point totals scored at runningback each week. This line is what you could have done if you were PERFECT at predicting the ranks of RB's each week. The black line is the PPG of a specific player, ranked accordingly. So the 1st black dot is the Arian Foster dot, 2nd is Darren Mcfadden etc. Lastly, the Red line is the average actual PPG of FBG's projections ranked accordingly. So the 1st dot on the red line corresponds to an average of all of FBG's highest rated RB for each week (This ended up being a frankenstein of lots of Foster and Frank Gore as I recall).
http://www.mediafire.com/?1y4vy01949bq9h7
Whats neat about that is the remarkable similarity between The season net PPG for RB's and the average of FBG's weekly projections. Basically if you had listened to FBG's each week and played their highest projected RB you would have gotten something almost identical to Arian Foster! This may not seem like a big deal, but don't ignore that one is a projection and one is hindsight. Clearly there is still room for improvement in projections as I want to get to that illusive green line, and make kill people and not go to jail money.
I'll update further if I encounter anything particularly interesting as I look into the data further.
Perhaps more clearly, essentially 60% of the variability in players actual points per game can be explained by footballguys pre-season projections! I have done no comparison across other websites but that is extremely impressive to me as someone who has examined a great deal of data types.
http://www.mediafire.com/?c9tgjnplgmklxj1 - All player projections vs actual results possible using FBG Data dominator (QB=32 players, RB=123 players, WR= 136 players, TE= 81 players
http://www.mediafire.com/?142p1z4lwqnqc2x - Same format of graph looking only at top starters (QB=12, RB=24, WR=36,TE=12).
However, The in season projections are perhaps less promising in usefulness. The following are a pooled data set of all projection-actual pairs for wr/rb/te (note no qb's) using dodds weekly projections. To read the following graph projection value of 5 means all projections between 5.00 and 5.99 etc. The number of projections is the number of times Dodds projected a player to score in that range on the season. It's a graph of median values (50% of actual below and 50% above), and the bars around each point are the Inner quartile range (25 percentile to 75th percentile). In essence 50% of all actual scores were between those ranges (Yes the ranges are that big!).
http://www.mediafire.com/?dwptr8t0o8ldnas
The means (not displayed on graph) of the lower projection values ( about 1 thru 10) are quite accurate as you can see by the above table. Which is to say if Dodds predicts a player will score in the 6.00-6.99 range the mean value of actual responses was 6.37 for example. In that projection range the actual values not only keep increasing, but they are accurate to the projected values themselves. Once you reach projections higher than 10 however, the data becomes more erratic. This is possible due to the much smaller sample sizes involved but nonetheless limits the usefulness of discerning between players projected at more than 10 points in a given week.
Another thing to be considered is that this data is a pooled grouping of 3 different positions (wr te and rb) the relationships above may be more or less accurate depending on position.
I think the following graph does a good job of showing the value of the projections (for RB's at least, more to follow). The legend might be a little confusing. Basically its 3 lines that show the following information: The green line is the season average of PPG at a given rank. So the 1st rank is the average of all of the highest point totals scored at runningback each week. This line is what you could have done if you were PERFECT at predicting the ranks of RB's each week. The black line is the PPG of a specific player, ranked accordingly. So the 1st black dot is the Arian Foster dot, 2nd is Darren Mcfadden etc. Lastly, the Red line is the average actual PPG of FBG's projections ranked accordingly. So the 1st dot on the red line corresponds to an average of all of FBG's highest rated RB for each week (This ended up being a frankenstein of lots of Foster and Frank Gore as I recall).
http://www.mediafire.com/?1y4vy01949bq9h7
Whats neat about that is the remarkable similarity between The season net PPG for RB's and the average of FBG's weekly projections. Basically if you had listened to FBG's each week and played their highest projected RB you would have gotten something almost identical to Arian Foster! This may not seem like a big deal, but don't ignore that one is a projection and one is hindsight. Clearly there is still room for improvement in projections as I want to get to that illusive green line, and make kill people and not go to jail money.
I'll update further if I encounter anything particularly interesting as I look into the data further.
Last edited by a moderator: