AEnigma wrote:lessthanjake wrote:AEnigma wrote:Because I am not in the business of compiling lists of outputs for the purpose of advancing a chosen fan narrative. You see a metric’s
creator say, “My metric rates this player worse when I add box components,” and rather than engage with that concept, your reaction is to dump a compilation of irrelevant other metrics in a transparent slander attempt. That is pure ideology, but characteristically, you prefer to pretend it is actually something everyone
else does.

I hope you realize that compiling all available data is virtually definitionally the opposite of “advancing a chosen fan narrative.”
If you did it comprehensively and neutrally, sure. But time and time again, you act selectively and emphasise what is most favourable to the narrative you want to push.
Lol, okay. This is just a really odd response to me, given that I’m more insistent than probably anyone on these forums that we should look at *all* available data. The “compilations” of data I’ve provided have often included individual metrics that, taken by themselves, would *not* support the position I’m arguing. But the data picture as a whole does, and that is what I argue is most important, especially given that individual metrics will have various forms of error and flaws. If the standard is that I must be aware of and immediately recall every single data point in existence before compiling data in a post, then I’m sure I’ve not always met that (nor would that be reasonable to expect). And you’re more than free to point out data points I’ve missed, anytime I’ve compiled data. But the idea that I’ve systematically excluded data points that go against my views is actually demonstrably untrue. And I think you probably know that.
Meanwhile, it is people you typically agree with (and posts that you And-1) that explicitly argue that I’m wrong to compile data and to look at that entire picture. They instead argue that we should look at only information that conveniently has results they like the most—accompanied by some ham-fisted argument about why that piece of information is the only reliable data. It is very clearly not me that “selectively . . . emphazise[s] what is most favorable to the narrative” being pushed. People you agree with very explicitly do that, and I very explicitly push back on that approach, with a lot of ink having been spilt on many threads on these forums about exactly this disagreement in terms of approach. You truly are barking up the wrong tree.
rather than objecting to the presentation of fulsome information and/or retreating to the use of a particular metric or two that gives you output you like.
I thought you said you were being comprehensive?

What is your point? Me presenting fulsome information that, taken as a whole, leads to a particular conclusion is really not mutually exclusive with the existence of individual data points that are less favorable to that conclusion. That’s the whole point of looking at a compilation of data rather than just one or two individual data points! If every data point always said the same thing, then it wouldn’t be necessary!
I have repeatedly said I do not care much about all-in-ones, although not in the sense that I see them all equally. My issue is on the emphasis and on the conflation of them with some idea of objective player quality when they are all still highly dependent on circumstance (as I have repeatedly gone over).
Agreed that these metrics are dependent on circumstances. This is something I’m more than happy to talk about in any individual instance. And, indeed, I’ve talked about it many times before. For instance, I’ve talked about how, in terms of impact data, it is more favorable for a star player if the team’s offensive system and roster is specifically built to maximize him, rather than it being built to maximize the rest of the roster while assuming that the star player will still eat regardless. I think we should all probably be able to agree with that general statement, even if we might disagree on exactly how that concept applies to specific circumstances. I’ve also talked about how I think that, in a vacuum, being on a worse team may be better for purposes of impact data, because even if you control for how good a player’s teammates are, a player will actually have more effect on a bad team than a good team (because they’re filling in more gaping holes), and impact data is essentially trying to isolate out the impact of that individual player. That’s a conclusion that I’m less certain about but it makes intuitive sense to me. I’ve also talked a lot about the mental effect of teams not really trying to contend, as well as the effect on impact of things like the rubber-band effect. I could go on. I’ve talked very extensively on these forums about circumstances affecting the output of impact data.
These are all interesting topics, and there’s surely more contextual issues than those. You’ll find that I’ve been very consistent in saying that impact data is all flawed and prone to error (something that people you consistently agree with have somehow taken issue with). The fact that I’ve compiled large amounts of data reflects the fact that I think a compilation of data gives a data picture that minimizes flaws and error in the data as much as possible. But I’ve definitely never said data is perfect or all we should look at. Very much to the contrary! There may be contextual factors that affect *all* impact data, which is something that looking at all data would not account for.
To tie things back a bit, if you want to respond to a compilation of data I provide by arguing that that data goes a certain direction due to contextual factors that essentially bias all the data, then that would probably be a good discussion! Unfortunately, that’s very rarely how you or others react to compilations of data. The most common reaction is instead to say that we should only cherry-pick out certain specific pieces of data instead of looking at the whole picture, or to quibble with the compilation in some trivial way.
As for the fact that the MAMBA creator says that LeBron gets worse in his metric the more you weight the prior, I have no doubt that that’s true, but I find it extremely odd that you suggest I did not “engage with that concept.” I engaged with it by pointing out that, given that we have lots of RAPM output that is not consistent with that (at least as it relates to the specific topics you and I have discussed), that seems to just be a relic of the specific type of RAPM that the creator is using. This seems obviously true. If we have a bunch of multi-year RAPM data sets and MAMBA tends to be a little bit more favorable to LeBron than that available RAPM data, then it is pretty obviously the case that the only way the MAMBA prior pulls LeBron down from RAPM more than anyone else is if the RAPM that MAMBA uses is much more favorable to LeBron than a bunch of other RAPM data we have. So the MAMBA creator’s statement basically just means that the RAPM methodology he uses is very high on LeBron compared to other RAPM. And that’s a good fact for LeBron (though we don’t know how good, since I don’t think we actually have just the RAPM, and he doesn’t look the best once we add the prior). We don’t actually know what of the many versions of RAPM is the best, though.
And why do you think he — and indeed all these metrics, regardless of their actual “preference” — would use a different form of RAPM from what you cite from github or thebasketballdatabase?
There are plenty of logical reasons to prefer one type of RAPM over another, because there are a lot of methodological decisions involved. For instance, do we apply a minutes cutoff for players who didn’t play much (and if so, how much of one)? Do we apply a luck adjustment, and if so, what are we assuming is simply the product of luck? Do we apply an adjustment for the rubber band effect? Do we include some kind of basic prior? And, if we do include a basic prior, what is it based on? Is it based on minutes played? Is the prior based on the prior years’ RAPM? If it’s based on the prior years’ RAPM, how are you weighing each year? Are years that are further back given less weight, and what are those weights and how far are you going back? There’s of course a lot of these decisions. It is in no way clear what is the “right” answer on these methodological questions. And that’s the point! Assuming that one particular form of RAPM is the objective truth is definitely pretty tenuous, especially as it relates to a conclusion that is inconsistent with a lot of other forms of RAPM.