The voting process for the NBAâs most prestigious awardsâsuch as MVP, DPOY, and its All-NBA Teamsâhas always been fairly ill-defined in terms of both who votes and how they voted. The rules have changed drastically over the years and the full voting results werenât even published until the 2015 season.
Since that 2015 season, the NBA has released the results as PDF-formatted documents tabulated by Ernst & Young LLP. And while this was certainly a major step forward in transparency, the format isnât very useful for data analysis.
Iâm working on changing that.
In 2016, the NBA players and Board of Governors ratified a new Collective Bargaining Agreement. This agreement included a âdesignated playerâ exception that allows a team to sign one of its own players to a five-year maximum contract extension, according to the following criteria:
- He makes one of the three all-NBA teams or is named either defensive player of the year or most valuable player the previous season.
- He has made one of the three all-NBA teams or has been named defensive player of the year in two of the prior three seasons or the leagueâs most valuable player in one of the three prior seasons.
This exception comes in addition to the well-known âDerrick Rose Rule,â which incentivizes making All-NBA teams during a playerâs first four years. In response to the heightened stakes of its media-based awards, the NBA also made a few changes to its voter-selection process:
- Decreased the number of eligible voters for each award from 130 to 100.
- Limited the selection pool to âindependentâ media members (no radio/television broadcasters or writers associated with a particular team).
- Required at least one voter per NBA market.
In total, the NBAâs media-based awards have more meaning than ever and understanding the process has become all the more important.
The NBA has 9 distinct media-chosen awards, each with its own number
of placements and scoring system (more on that later). The table below
summarizes each voterâs ballot for a given awardâMost Valuable
Player (MVP
), Coach of the Year (COY
), Rookie of the Year (ROY
),
Defensive Player of the Year (DPOY
), Most Improved Player (MIP
),
6th Man of the Year (6th
), All-NBA (All-NBA
), All-Defense
(All-Defensive
), and All-Rookie (All-Rookie
).
The ultimate goal of this project is to provide a means of assessing the quality of a given ballot. A common complaint with the existing process is that voters are somehow âbiasedâ or are actively supporting a certain ânarrative.â Unfortunately, although itâs straightforward to describe the perceived problem, itâs much harder to actually identify it in practice.
For example, one of the most high-profile cases in recent times was Gary Washburnâs decision to select Carmelo Anthony over LeBron James as the 2013 MVP, effectively robbing him of the chance to become the NBAâs first unanimous choice (a feat Stephen Curry would later accomplish in 2016).
âI was heated,â James told Chris Haynes, then of Cleveland.com. âBut I knew all along [I wasnât getting a unanimous vote]. I just knew it, man.â
â A brief history of LeBron James disagreeing with awards voters
While you might be tempted to say that identifying unusual ballots (such as Washburnâs) is a good indication of poor choices, itâs really not that simpleâindeed, what if itâs the consensus itself thatâs âwrongâ?
This was exactly the case in 2021, according to Jayson Tatum:
âI know I should have made it with the season I had,â Tatum told host Ashley Nevel. âI mean $33 million on the line. Obviously, that would make anyone feel some type of way. And I wasnât necessarily upset about losing the money. I think I just felt like the way I was playing, everything I did, I thought it should have been a no-brainer. I think I was just more frustrated with that.â
However, even if it may not be possible to derive objective conclusions from an inherently subjective process, we can still perform some interesting data analysis.
The first type of analysis weâll perform is pretty standard in the fields of data science and statistics: the search for outliers in our data set.
But what exactly constitutes an outlier?
To answer this question, we must first understand the two types of
ballots: there are ranked lists (MVP
, COY
, DPOY
, âŚ) and team
selections (All-NBA
, All-Rookie
, and All-Defensive
).
For ranked-list awards, we define outliers as ballots that are the most dissimilar to the final result (the consensus). To help us do this, weâre going to use the Rank-Biased Overlap (RBO) metric.
The RBO metric is a bounded ([0, 1]
) similarity measure that includes
top-weightedness: (dis)agreements at the top of two lists will
weigh more heavily than the same (dis)agreements towards the bottom
of the lists.
Below is a visualization of all RBO values for each ranked award.
During the same interview in which Jayson Tatum expressed his displeasure with the 2020-21 All-NBA 3rd Team voting results, he also alluded to the need for a more objective criteria for making these All-NBA teams:
âI think what they do need to change is â itâs kind of opinion-based,â he explained. â100 media members have the vote, and whatâs the criteria, right? Is there a certain amount of games you need to play. Should you be in playoff contention? âŚâ
While the NBA is unlikely to implement such a criteria any time soon, the idea leads to an interesting question: Is there an implicit criteria that the media follows? We can begin to answer this by constructing a profile of what it takesâhistoricallyâto make an All-NBA team.