This is part 5 in my series on writing modern idiomatic pandas.


Reshaping & Tidy Data

Structuring datasets to facilitate analysis (Wickham 2014)

So, you’ve sat down to analyze a new dataset. What do you do first?

In episode 11 of Not So Standard Deviations, Hilary and Roger discussed their typical approaches. I’m with Hilary on this one, you should make sure your data is tidy. Before you do any plots, filtering, transformations, summary statistics, regressions… Without a tidy dataset, you’ll be fighting your tools to get the result you need. With a tidy dataset, it’s relatively easy to do all of those.

Hadley Wickham kindly summarized tidiness as a dataset where

  1. Each variable forms a column
  2. Each observation forms a row
  3. Each type of observational unit forms a table

And today we’ll only concern ourselves with the first two. As quoted at the top, this really is about facilitating analysis: going as quickly as possible from question to answer.

%matplotlib inline

import os
import numpy as np
import pandas as pd
import seaborn as sns
import matplotlib.pyplot as plt

if int(os.environ.get("MODERN_PANDAS_EPUB", 0)):
    import prep # noqa

pd.options.display.max_rows = 10
sns.set(style='ticks', context='talk')

NBA Data

This StackOverflow question asked about calculating the number of days of rest NBA teams have between games. The answer would have been difficult to compute with the raw data. After transforming the dataset to be tidy, we’re able to quickly get the answer.

We’ll grab some NBA game data from basketball-reference.com using pandas’ read_html function, which returns a list of DataFrames.

fp = 'data/nba.csv'

if not os.path.exists(fp):
    tables = pd.read_html("http://www.basketball-reference.com/leagues/NBA_2016_games.html")
    games = tables[0]
    games.to_csv(fp)
else:
    games = pd.read_csv(fp)
games.head()

DateStart (ET)Unnamed: 2Visitor/NeutralPTSHome/NeutralPTS.1Unnamed: 7Notes
0OctoberNaNNaNNaNNaNNaNNaNNaNNaN
1Tue, Oct 27, 20158:00 pmBox ScoreDetroit Pistons106.0Atlanta Hawks94.0NaNNaN
2Tue, Oct 27, 20158:00 pmBox ScoreCleveland Cavaliers95.0Chicago Bulls97.0NaNNaN
3Tue, Oct 27, 201510:30 pmBox ScoreNew Orleans Pelicans95.0Golden State Warriors111.0NaNNaN
4Wed, Oct 28, 20157:30 pmBox ScorePhiladelphia 76ers95.0Boston Celtics112.0NaNNaN

Side note: pandas’ read_html is pretty good. On simple websites it almost always works. It provides a couple parameters for controlling what gets selected from the webpage if the defaults fail. I’ll always use it first, before moving on to BeautifulSoup or lxml if the page is more complicated.

As you can see, we have a bit of general munging to do before tidying. Each month slips in an extra row of mostly NaNs, the column names aren’t too useful, and we have some dtypes to fix up.

column_names = {'Date': 'date', 'Start (ET)': 'start',
                'Unamed: 2': 'box', 'Visitor/Neutral': 'away_team', 
                'PTS': 'away_points', 'Home/Neutral': 'home_team',
                'PTS.1': 'home_points', 'Unamed: 7': 'n_ot'}

games = (games.rename(columns=column_names)
    .dropna(thresh=4)
    [['date', 'away_team', 'away_points', 'home_team', 'home_points']]
    .assign(date=lambda x: pd.to_datetime(x['date'], format='%a, %b %d, %Y'))
    .set_index('date', append=True)
    .rename_axis(["game_id", "date"])
    .sort_index())
games.head()

away_teamaway_pointshome_teamhome_points
game_iddate
12015-10-27Detroit Pistons106.0Atlanta Hawks94.0
22015-10-27Cleveland Cavaliers95.0Chicago Bulls97.0
32015-10-27New Orleans Pelicans95.0Golden State Warriors111.0
42015-10-28Philadelphia 76ers95.0Boston Celtics112.0
52015-10-28Chicago Bulls115.0Brooklyn Nets100.0

A quick aside on that last block.

  • dropna has a thresh argument. If at least thresh items are missing, the row is dropped. We used it to remove the “Month headers” that slipped into the table.
  • assign can take a callable. This lets us refer to the DataFrame in the previous step of the chain. Otherwise we would have to assign temp_df = games.dropna()... And then do the pd.to_datetime on that.
  • set_index has an append keyword. We keep the original index around since it will be our unique identifier per game.
  • We use .rename_axis to set the index names (this behavior is new in pandas 0.18; before .rename_axis only took a mapping for changing labels).

The Question:

How many days of rest did each team get between each game?

Whether or not your dataset is tidy depends on your question. Given our question, what is an observation?

In this case, an observation is a (team, game) pair, which we don’t have yet. Rather, we have two observations per row, one for home and one for away. We’ll fix that with pd.melt.

pd.melt works by taking observations that are spread across columns (away_team, home_team), and melting them down into one column with multiple rows. However, we don’t want to lose the metadata (like game_id and date) that is shared between the observations. By including those columns as id_vars, the values will be repeated as many times as needed to stay with their observations.

tidy = pd.melt(games.reset_index(),
               id_vars=['game_id', 'date'], value_vars=['away_team', 'home_team'],
               value_name='team')
tidy.head()

game_iddatevariableteam
012015-10-27away_teamDetroit Pistons
122015-10-27away_teamCleveland Cavaliers
232015-10-27away_teamNew Orleans Pelicans
342015-10-28away_teamPhiladelphia 76ers
452015-10-28away_teamChicago Bulls

The DataFrame tidy meets our rules for tidiness: each variable is in a column, and each observation (team, date pair) is on its own row. Now the translation from question (“How many days of rest between games”) to operation (“date of today’s game - date of previous game - 1”) is direct:

# For each team... get number of days between games
tidy.groupby('team')['date'].diff().dt.days - 1
0       NaN
1       NaN
2       NaN
3       NaN
4       NaN
       ... 
2455    7.0
2456    1.0
2457    1.0
2458    3.0
2459    2.0
Name: date, Length: 2460, dtype: float64

That’s the essence of tidy data, the reason why it’s worth considering what shape your data should be in. It’s about setting yourself up for success so that the answers naturally flow from the data (just kidding, it’s usually still difficult. But hopefully less so).

Let’s assign that back into our DataFrame

tidy['rest'] = tidy.sort_values('date').groupby('team').date.diff().dt.days - 1
tidy.dropna().head()

game_iddatevariableteamrest
452015-10-28away_teamChicago Bulls0.0
892015-10-28away_teamCleveland Cavaliers0.0
14152015-10-28away_teamNew Orleans Pelicans0.0
17182015-10-29away_teamMemphis Grizzlies0.0
18192015-10-29away_teamDallas Mavericks0.0

To show the inverse of melt, let’s take rest values we just calculated and place them back in the original DataFrame with a pivot_table.

by_game = (pd.pivot_table(tidy, values='rest',
                          index=['game_id', 'date'],
                          columns='variable')
             .rename(columns={'away_team': 'away_rest',
                              'home_team': 'home_rest'}))
df = pd.concat([games, by_game], axis=1)
df.dropna().head()

away_teamaway_pointshome_teamhome_pointsaway_resthome_rest
game_iddate
182015-10-29Memphis Grizzlies112.0Indiana Pacers103.00.00.0
192015-10-29Dallas Mavericks88.0Los Angeles Clippers104.00.00.0
202015-10-29Atlanta Hawks112.0New York Knicks101.01.00.0
212015-10-30Charlotte Hornets94.0Atlanta Hawks97.01.00.0
222015-10-30Toronto Raptors113.0Boston Celtics103.01.01.0

One somewhat subtle point: an “observation” depends on the question being asked. So really, we have two tidy datasets, tidy for answering team-level questions, and df for answering game-level questions.

One potentially interesting question is “what was each team’s average days of rest, at home and on the road?” With a tidy dataset (the DataFrame tidy, since it’s team-level), seaborn makes this easy (more on seaborn in a future post):

sns.set(style='ticks', context='paper')
g = sns.FacetGrid(tidy, col='team', col_wrap=6, hue='team', size=2)
g.map(sns.barplot, 'variable', 'rest');

png

An example of a game-level statistic is the distribution of rest differences in games:

df['home_win'] = df['home_points'] > df['away_points']
df['rest_spread'] = df['home_rest'] - df['away_rest']
df.dropna().head()

away_teamaway_pointshome_teamhome_pointsaway_resthome_resthome_winrest_spread
game_iddate
182015-10-29Memphis Grizzlies112.0Indiana Pacers103.00.00.0False0.0
192015-10-29Dallas Mavericks88.0Los Angeles Clippers104.00.00.0True0.0
202015-10-29Atlanta Hawks112.0New York Knicks101.01.00.0False-1.0
212015-10-30Charlotte Hornets94.0Atlanta Hawks97.01.00.0True-1.0
222015-10-30Toronto Raptors113.0Boston Celtics103.01.01.0False0.0
delta = (by_game.home_rest - by_game.away_rest).dropna().astype(int)
ax = (delta.value_counts()
    .reindex(np.arange(delta.min(), delta.max() + 1), fill_value=0)
    .sort_index()
    .plot(kind='bar', color='k', width=.9, rot=0, figsize=(12, 6))
)
sns.despine()
ax.set(xlabel='Difference in Rest (Home - Away)', ylabel='Games');

png

Or the win percent by rest difference

fig, ax = plt.subplots(figsize=(12, 6))
sns.barplot(x='rest_spread', y='home_win', data=df.query('-3 <= rest_spread <= 3'),
            color='#4c72b0', ax=ax)
sns.despine()

png

Stack / Unstack

Pandas has two useful methods for quickly converting from wide to long format (stack) and long to wide (unstack).

rest = (tidy.groupby(['date', 'variable'])
            .rest.mean()
            .dropna())
rest.head()
date        variable 
2015-10-28  away_team    0.000000
            home_team    0.000000
2015-10-29  away_team    0.333333
            home_team    0.000000
2015-10-30  away_team    1.083333
Name: rest, dtype: float64

rest is in a “long” form since we have a single column of data, with multiple “columns” of metadata (in the MultiIndex). We use .unstack to move from long to wide.

rest.unstack().head()

variableaway_teamhome_team
date
2015-10-280.0000000.000000
2015-10-290.3333330.000000
2015-10-301.0833330.916667
2015-10-310.1666670.833333
2015-11-011.1428571.000000

unstack moves a level of a MultiIndex (innermost by default) up to the columns. stack is the inverse.

rest.unstack().stack()
date        variable 
2015-10-28  away_team    0.000000
            home_team    0.000000
2015-10-29  away_team    0.333333
            home_team    0.000000
2015-10-30  away_team    1.083333
                           ...   
2016-04-11  home_team    0.666667
2016-04-12  away_team    1.000000
            home_team    1.400000
2016-04-13  away_team    0.500000
            home_team    1.214286
Length: 320, dtype: float64

With .unstack you can move between those APIs that expect there data in long-format and those APIs that work with wide-format data. For example, DataFrame.plot(), works with wide-form data, one line per column.

with sns.color_palette() as pal:
    b, g = pal.as_hex()[:2]

ax=(rest.unstack()
        .query('away_team < 7')
        .rolling(7)
        .mean()
        .plot(figsize=(12, 6), linewidth=3, legend=False))
ax.set(ylabel='Rest (7 day MA)')
ax.annotate("Home", (rest.index[-1][0], 1.02), color=g, size=14)
ax.annotate("Away", (rest.index[-1][0], 0.82), color=b, size=14)
sns.despine()

png

The most conenient form will depend on exactly what you’re doing. When interacting with databases you’ll often deal with long form data. Pandas’ DataFrame.plot often expects wide-form data, while seaborn often expect long-form data. Regressions will expect wide-form data. Either way, it’s good to be comfortable with stack and unstack (and MultiIndexes) to quickly move between the two.

Mini Project: Home Court Advantage?

We’ve gone to all that work tidying our dataset, let’s put it to use. What’s the effect (in terms of probability to win) of being the home team?

Step 1: Create an outcome variable

We need to create an indicator for whether the home team won. Add it as a column called home_win in games.

df['home_win'] = df.home_points > df.away_points

Step 2: Find the win percent for each team

In the 10-minute literature review I did on the topic, it seems like people include a team-strength variable in their regressions. I suppose that makes sense; if stronger teams happened to play against weaker teams at home more often than away, it’d look like the home-effect is stronger than it actually is. We’ll do a terrible job of controlling for team strength by calculating each team’s win percent and using that as a predictor. It’d be better to use some kind of independent measure of team strength, but this will do for now.

We’ll use a similar melt operation as earlier, only now with the home_win variable we just created.

wins = (
    pd.melt(df.reset_index(),
            id_vars=['game_id', 'date', 'home_win'],
            value_name='team', var_name='is_home',
            value_vars=['home_team', 'away_team'])
   .assign(win=lambda x: x.home_win == (x.is_home == 'home_team'))
   .groupby(['team', 'is_home'])
   .win
   .agg(['sum', 'count', 'mean'])
   .rename(columns=dict(sum='n_wins',
                        count='n_games',
                        mean='win_pct'))
)
wins.head()

n_winsn_gameswin_pct
teamis_home
Atlanta Hawksaway_team21.0410.512195
home_team27.0410.658537
Boston Celticsaway_team20.0410.487805
home_team28.0410.682927
Brooklyn Netsaway_team7.0410.170732

Pause for visualiztion, because why not

g = sns.FacetGrid(wins.reset_index(), hue='team', size=7, aspect=.5, palette=['k'])
g.map(sns.pointplot, 'is_home', 'win_pct').set(ylim=(0, 1));

png

(It’d be great if there was a library built on top of matplotlib that auto-labeled each point decently well. Apparently this is a difficult problem to do in general).

g = sns.FacetGrid(wins.reset_index(), col='team', hue='team', col_wrap=5, size=2)
g.map(sns.pointplot, 'is_home', 'win_pct')
<seaborn.axisgrid.FacetGrid at 0x11a0fe588>

png

Those two graphs show that most teams have a higher win-percent at home than away. So we can continue to investigate. Let’s aggregate over home / away to get an overall win percent per team.

win_percent = (
    # Use sum(games) / sum(games) instead of mean
    # since I don't know if teams play the same
    # number of games at home as away
    wins.groupby(level='team', as_index=True)
        .apply(lambda x: x.n_wins.sum() / x.n_games.sum())
)
win_percent.head()
team
Atlanta Hawks        0.585366
Boston Celtics       0.585366
Brooklyn Nets        0.256098
Charlotte Hornets    0.585366
Chicago Bulls        0.512195
dtype: float64
win_percent.sort_values().plot.barh(figsize=(6, 12), width=.85, color='k')
plt.tight_layout()
sns.despine()
plt.xlabel("Win Percent")

png

Is there a relationship between overall team strength and their home-court advantage?

plt.figure(figsize=(8, 5))
(wins.win_pct
    .unstack()
    .assign(**{'Home Win % - Away %': lambda x: x.home_team - x.away_team,
               'Overall %': lambda x: (x.home_team + x.away_team) / 2})
     .pipe((sns.regplot, 'data'), x='Overall %', y='Home Win % - Away %')
)
sns.despine()
plt.tight_layout()

png

Let’s get the team strength back into df. You could you pd.merge, but I prefer .map when joining a Series.

df = df.assign(away_strength=df['away_team'].map(win_percent),
               home_strength=df['home_team'].map(win_percent),
               point_diff=df['home_points'] - df['away_points'],
               rest_diff=df['home_rest'] - df['away_rest'])
df.head()

away_teamaway_pointshome_teamhome_pointsaway_resthome_resthome_winrest_spreadaway_strengthhome_strengthpoint_diffrest_diff
game_iddate
12015-10-27Detroit Pistons106.0Atlanta Hawks94.0NaNNaNFalseNaN0.5365850.585366-12.0NaN
22015-10-27Cleveland Cavaliers95.0Chicago Bulls97.0NaNNaNTrueNaN0.6951220.5121952.0NaN
32015-10-27New Orleans Pelicans95.0Golden State Warriors111.0NaNNaNTrueNaN0.3658540.89024416.0NaN
42015-10-28Philadelphia 76ers95.0Boston Celtics112.0NaNNaNTrueNaN0.1219510.58536617.0NaN
52015-10-28Chicago Bulls115.0Brooklyn Nets100.00.0NaNFalseNaN0.5121950.256098-15.0NaN
import statsmodels.formula.api as sm

df['home_win'] = df.home_win.astype(int)  # for statsmodels
mod = sm.logit('home_win ~ home_strength + away_strength + home_rest + away_rest', df)
res = mod.fit()
res.summary()
Optimization terminated successfully.
         Current function value: 0.552792
         Iterations 6
Logit Regression Results
Dep. Variable:home_winNo. Observations:1213
Model:LogitDf Residuals:1208
Method:MLEDf Model:4
Date:Sun, 03 Sep 2017Pseudo R-squ.:0.1832
Time:07:25:41Log-Likelihood:-670.54
converged:TrueLL-Null:-820.91
LLR p-value:7.479e-64
coefstd errzP>|z|[0.0250.975]
Intercept0.07070.3140.2250.822-0.5460.687
home_strength5.42040.46511.6470.0004.5086.333
away_strength-4.74450.452-10.5060.000-5.630-3.859
home_rest0.08940.0791.1370.255-0.0650.243
away_rest-0.04220.067-0.6290.529-0.1740.089

The strength variables both have large coefficeints (really we should be using some independent measure of team strength here, win_percent is showing up on the left and right side of the equation). The rest variables don’t seem to matter as much.

With .assign we can quickly explore variations in formula.

(sm.Logit.from_formula('home_win ~ strength_diff + rest_spread',
                       df.assign(strength_diff=df.home_strength - df.away_strength))
    .fit().summary())
Optimization terminated successfully.
         Current function value: 0.553499
         Iterations 6
Logit Regression Results
Dep. Variable:home_winNo. Observations:1213
Model:LogitDf Residuals:1210
Method:MLEDf Model:2
Date:Sun, 03 Sep 2017Pseudo R-squ.:0.1821
Time:07:25:41Log-Likelihood:-671.39
converged:TrueLL-Null:-820.91
LLR p-value:1.165e-65
coefstd errzP>|z|[0.0250.975]
Intercept0.46100.0686.7560.0000.3270.595
strength_diff5.06710.34914.5210.0004.3835.751
rest_spread0.05660.0620.9120.362-0.0650.178
mod = sm.Logit.from_formula('home_win ~ home_rest + away_rest', df)
res = mod.fit()
res.summary()
Optimization terminated successfully.
         Current function value: 0.676549
         Iterations 4
Logit Regression Results
Dep. Variable:home_winNo. Observations:1213
Model:LogitDf Residuals:1210
Method:MLEDf Model:2
Date:Sun, 03 Sep 2017Pseudo R-squ.:0.0003107
Time:07:25:41Log-Likelihood:-820.65
converged:TrueLL-Null:-820.91
LLR p-value:0.7749
coefstd errzP>|z|[0.0250.975]
Intercept0.36670.0943.8890.0000.1820.552
home_rest0.03380.0690.4860.627-0.1020.170
away_rest-0.04200.061-0.6930.488-0.1610.077

Overall not seeing to much support for rest mattering, but we got to see some more tidy data.

That’s it for today. Next time we’ll look at data visualization.