Dr Positivity wrote:I find it frustrating how hard it is to find "honest" APM, meaning "+/- adjusted by everyone else's" and that's IT, let the results be what they may. Of course the reason people pushed past that is the more clean APM likely had crazy looking or non predictive results. So the response has been to try and push the ball past the goal line to predictive soundness with progressively sketchier fudging. And this was true before RPM or xRAPM. Tbh I never really trusted APM after reading this
http://arturogalletti.wordpress.com/201 ... g-a-model/ And as importantly, that no APBR person in the comment section found a way to debunk the claims or deny that Step 3 is happening and destroys the stat
So you know Got Buckets now delivers an APM (and mystic says good things about their source):
http://www.gotbuckets.com/statistics/apm/2014-apm/I"m with you that it's been awful frustrating seeing data disappear. Hence my hissy fits.
Re: Arturo & APM. Dude has some serious misunderstanding about the stat. I'll say a bit here. Feel free to ask more:
When you refer to Step 3, I assume you mean this:
Step 3 :Calculating Adjusted +/-
The final step is to take the Pure regression and the Stats model and adds them up by player like so:
APM = x* Pure +/- + (1-x)*Statistical +/-
And proceed to adjust x between 10% and 90% for each player to minimize the error. In essence he tweaks the rating to get a high R-Square.
To summarize, the APM model calculates two variables with a low correlation to wins (R^2 <5%) and adds them up to minimize the error and guarantee a 90%+ Rsq. for the overall model.
Funny that.
The biggest issue with this is that he's writing in 2011 using a source from 2004 as gospel, when in fact no one sense that source did the thing that Arturo is objecting to - and we were years into the RAPM era by 2011 anyway. I'm not going to sort through the comments on the page years after the fact, but I can tell you that people marveled at how either clueless or how disingenuous he was. Why?
Because there was a history between the Wages of Wins (WoW) people and APBRmetrics people where APBRMetrics people basically said "APBRmetrics is where basketball stat people talk about basketball stats online, come talk with us over there", and the WoW people didn't. The APBRmetric people kept trying to facilitate the discussion by commenting on the WoW boards, and many got their comments deleted and their accounts banned. I witnessed it personally, and their behavior in no way warranted such behavior - and I say this speaking as a long term mod of the strictest basketball board around.
So then after all that, Arturo decides to look into this "enemy" stat, and he completely wastes his time precisely because he chose to try to figure it out on his own, taking every opportunity to just criticize rather than ask experts.
As for why the 2004 source (Roesenbaum) did what he did: The concern with +/- is always the noise. Results are not as consistent with this stat as they are with box score metrics. Rosenbaum as a guy just jumping in and playing around had a couple ideas in his head (APM & SPM), and just decided to do both, and then combine them undoubtedly thinking that would give maybe some more reliable data. There's nothing wrong with any of this really, but that was basically it for Rosenbaum. He never published any more results, and it wasn't until other guys (Ilardi, Sill, etc) got involved that any of us started seriously using APM data.
So, what did it take for people to get "sold" on it? Speaking just for myself:
1) I had to figure out what the quirks were. If you just take a one-year APM value at face value for every player in the league you're going to have some crazy ideas in your head, so you would never do that. That doesn't mean though that things are purely random or that all the results are equally slippery.
Consider 2006 where the Pistons had extreme success playing their starting lineup in huge minutes. This causes a massive dose of the what's called multicollinearity and it makes it so the APM data for those guys taken out of context isn't necessarily useful. But why would paint the issues we see with the Pistons with the same brush as a team where you had a more varied distribution of minutes? You wouldn't, if you knew what you were doing.
When points like this get made in rebuttal to guys like Arturo, or to non-stats guys, people tend to make a two-pronged criticism that has always seemed contradictory to me: 1) You're using APM for everything!... 2) except when you just ignore it. Whenever I see a person use a hammer to hammer a nail and not use a hammer to brush his teeth, that tends to make me think that the person knows what he's doing, but in these circumstances explanations tend to get dismissed as excuses.
2) People started doing stuff to minimize the noise. They did multi-year model, they did collections of one year models, they created "regularized" APM, they varied the timespan for that too, then they started making regression models on things like rebounding, turnovers, etc too. With every step along the way we've just gotten better and better data and better and better at using the day.
Now, folks like Arturo will stay cling to the fact that there's more noise in RAPM than in something like WP, but it's crucial to understand precisely what that means. It means that they are criticizing a model for lacking in reliability - which is fine - while giving no clear argument that the alternative they champion offer similar validity.
Fundamentally, WP and RAPM literally do not measure the same ting. WP like any box score metric is biased toward particular skills and foci associated with particular stats, and typically that means a preference for players on the offensive side of the ball. This means your ideal WP by definition is not your ideal basketball player. This is indisputable despite the fact that Berri and his guys have often tried to dispute it. The box score doesn't track everything, and so you can only go so far with it. The bias of the box score makes removal of that bias impossible.
So then, exactly how valuable is it to be an extremely reliable - extremely consistent - stat that gives a biased representation of what it purports to measure? Answer is clearly "depends". It can still be quite useful, but the reliability isn't even necessarily a good thing.
Here's the image I often use to explain the importance of +/-:

A stat like Wins Produced is like the results on the left. A stat like +/- is one like the middle. Which is better? What does that even mean being "better"? Clearly neither is the result on the right, so you can't just use one or the other. You've got to use both. (Disclosure: Of course we don't actually use WP typically, we use other stats that do what WP does better, but if those other stats didn't exist, I'd use WP.)