Lisa Yan
Content by Lisa Yan, Will Fithian, Joseph Gonzalez, Deborah Nolan, Sam Lau
import numpy as np
import pandas as pd
import matplotlib.pyplot as plt
import seaborn as sns
#%matplotlib inline
plt.rcParams['figure.figsize'] = (12, 9)
sns.set()
sns.set_context('talk')
np.set_printoptions(threshold=20, precision=2, suppress=True)
pd.set_option('display.max_rows', 30)
pd.set_option('display.max_columns', None)
pd.set_option('display.precision', 2)
# This option stops scientific notation for pandas
pd.set_option('display.float_format', '{:.2f}'.format)
There are many file types for storing structured data: CSV, TSV, JSON, XML, ASCII, SAS...
TSV (Tab-Separated Values) files are very similar to CSVs, but now items are delimited by tabs.
Let's check out cdc_tuberculosis.tsv
, which is the same data but now in a TSV.
To the Jupyter view!
To the Python view!
Quick Python reminders:
print()
prints each string (including the newline), and an additional newline on top of that.repr()
function to return the raw string with all special characters:with open("data/cdc_tuberculosis.tsv", "r") as f:
i = 0
for row in f:
print(repr(row)) # print raw strings
i += 1
if i > 3:
break
'\tNo. of TB cases\t\t\tTB incidence\t\t\n' 'U.S. jurisdiction\t2019\t2020\t2021\t2019\t2020\t2021\n' 'Total\t"8,900"\t"7,173"\t"7,860"\t2.71\t2.16\t2.37\n' 'Alabama\t87\t72\t92\t1.77\t1.43\t1.83\n'
The pd.read_csv
function also reads in TSVs if we specify the delimiter with parameter sep='\t'
(documentation).
tuberculosis_df_tsv = pd.read_csv("data/cdc_tuberculosis.tsv", sep='\t')
tuberculosis_df_tsv.head()
Unnamed: 0 | No. of TB cases | Unnamed: 2 | Unnamed: 3 | TB incidence | Unnamed: 5 | Unnamed: 6 | |
---|---|---|---|---|---|---|---|
0 | U.S. jurisdiction | 2019 | 2020 | 2021 | 2019.00 | 2020.00 | 2021.00 |
1 | Total | 8,900 | 7,173 | 7,860 | 2.71 | 2.16 | 2.37 |
2 | Alabama | 87 | 72 | 92 | 1.77 | 1.43 | 1.83 |
3 | Alaska | 58 | 58 | 58 | 7.91 | 7.92 | 7.92 |
4 | Arizona | 183 | 136 | 129 | 2.51 | 1.89 | 1.77 |
Side note: there was a question last time on how pandas differentiates a comma delimiter vs. a comma within the field itself, e.g., 8,900
. Check out the documentation for the quotechar
parameter.
In the interest of reproducible data science we will download the data programatically. We have defined some helper functions in the ds100_utils.py file. I can then reuse these helper functions in many different notebooks.
# just run this cell
from ds100_utils import fetch_and_cache
Occasionally, you will want to modify code that you have imported from a local Python library. To reimport those modifications you can either use the python importlib library:
from importlib import reload
reload(utils)
or use iPython magic which will intelligently import code when files change:
%load_ext autoreload
%autoreload 2
covid_file = fetch_and_cache(
"https://data.cityofberkeley.info/api/views/xn6j-b766/rows.json?accessType=DOWNLOAD",
"confirmed-cases.json",
force=False)
covid_file # a file path wrapper object
Using cached version that was downloaded (UTC): Tue Jan 31 14:33:04 2023
PosixPath('data/confirmed-cases.json')
Often, I like to start my analysis by getting a rough estimate of the size of the data. This will help inform the tools I use and how I view the data. If it is relatively small I might use a text editor or a spreadsheet to look at the data. If it is larger, I might jump to more programmatic exploration or even used distributed computing tools.
However here we will use Python tools to probe the file.
Since these seem to be text files I might also want to investigate the number of lines, which often corresponds to the number of records.
import os
print(covid_file, "is", os.path.getsize(covid_file) / 1e6, "MB")
with open(covid_file, "r") as f:
print(covid_file, "is", sum(1 for l in f), "lines.")
data/confirmed-cases.json is 0.183341 MB data/confirmed-cases.json is 1559 lines.
Because we have a text file in a visual IDE like Jupyter/DataHub, I'm going to visually explore the data via the built-in file explorer.
with open(covid_file, "r") as f:
i = 0
for row in f:
print(repr(row)) # print raw strings
i += 1
if i > 5:
break
'{\n' ' "meta" : {\n' ' "view" : {\n' ' "id" : "xn6j-b766",\n' ' "name" : "COVID-19 Confirmed Cases",\n' ' "assetType" : "dataset",\n'
Back to the Python view.
In order to load the JSON file into pandas, Let's first do some EDA with the Python json
package to understand the particular structure of this JSON file so that we can decide what (if anything) to load into Pandas.
Python has relatively good support for JSON data since it closely matches the internal python object model. In the following cell we import the entire JSON datafile into a python dictionary using the json
package.
import json
with open(covid_file, "rb") as f:
covid_json = json.load(f)
The covid_json
variable is now a dictionary encoding the data in the file:
type(covid_json)
dict
We can list the keys to determine what data is stored in the object.
covid_json.keys()
dict_keys(['meta', 'data'])
Observation: The JSON dictionary contains a meta
key which likely refers to meta data (data about the data). Meta data often maintained with the data and can be a good source of additional information.
We can investigate the meta data further by examining the keys associated with the metadata.
covid_json['meta'].keys()
dict_keys(['view'])
The meta
key contains another dictionary called view
. This likely refers to meta-data about a particular "view" of some underlying database. We will learn more about views when we study SQL later in the class.
covid_json['meta']['view'].keys()
dict_keys(['id', 'name', 'assetType', 'attribution', 'averageRating', 'category', 'createdAt', 'description', 'displayType', 'downloadCount', 'hideFromCatalog', 'hideFromDataJson', 'newBackend', 'numberOfComments', 'oid', 'provenance', 'publicationAppendEnabled', 'publicationDate', 'publicationGroup', 'publicationStage', 'rowsUpdatedAt', 'rowsUpdatedBy', 'tableId', 'totalTimesRated', 'viewCount', 'viewLastModified', 'viewType', 'approvals', 'clientContext', 'columns', 'grants', 'metadata', 'owner', 'query', 'rights', 'tableAuthor', 'tags', 'flags'])
Notice that this a nested/recursive data structure. As we dig deeper we reveal more and more keys and the corresponding data:
meta
|-> data
| ... (haven't explored yet)
|-> view
| -> id
| -> name
| -> attribution
...
| -> description
...
| -> columns
...
There is a key called description in the view sub dictionary. This likely contains a description of the data:
print(covid_json['meta']['view']['description'])
Counts of confirmed COVID-19 cases among Berkeley residents by date. As of 6/21/22, this dataset will be updated weekly instead of daily. As of 11/14/22, this dataset only includes PCR cases.
We can look at a few entries in the data
field. This is what we'll load into Pandas.
for i in range(3):
print(f"{i:03} | {covid_json['data'][i]}")
000 | ['row-2yyp~r8a3~phgq', '00000000-0000-0000-F944-963C2BD56F87', 0, 1674521616, None, 1674521616, None, '{ }', '2019-12-01T00:00:00', '0', '0'] 001 | ['row-svsf_gzh2~cz9t', '00000000-0000-0000-BA8E-0D8297E66451', 0, 1674521616, None, 1674521616, None, '{ }', '2019-12-02T00:00:00', '0', '0'] 002 | ['row-w244_ivf6-rdcu', '00000000-0000-0000-A608-DB4DF9DB1B16', 0, 1674521616, None, 1674521616, None, '{ }', '2019-12-03T00:00:00', '0', '0']
Observations:
data
is a table!Back to the metadata.
Another potentially useful key in the metadata dictionary is the columns
. This returns a list:
type(covid_json['meta']['view']['columns'])
list
Let's go back to the file explorer.
Based on the contents of this key, what are reasonable names for each column in the data
table?
After our above EDA, let's finally go about loading the data (not the metadata) into a pandas dataframe.
In the following block of code we:
Translate the JSON records into a dataframe:
covid_json['meta']['view']['columns']
covid_json['data']
Remove columns that have no metadata description. This would be a bad idea in general but here we remove these columns since the above analysis suggests that they are unlikely to contain useful information.
tail
of the table.# Load the data from JSON and assign column titles
covid = pd.DataFrame(
covid_json['data'],
columns=[c['name'] for c in covid_json['meta']['view']['columns']])
covid.tail()
sid | id | position | created_at | created_meta | updated_at | updated_meta | meta | Date | New Cases | Cumulative Cases | |
---|---|---|---|---|---|---|---|---|---|---|---|
1144 | row-vvwh-m4y7.mr6x | 00000000-0000-0000-8DE9-36E51F431D12 | 0 | 1674521616 | None | 1674521616 | None | { } | 2023-01-18T00:00:00 | 6 | 22182 |
1145 | row-zex5.4t7g-ehhz | 00000000-0000-0000-7A55-165F57F862F6 | 0 | 1674521616 | None | 1674521616 | None | { } | 2023-01-19T00:00:00 | 0 | 22182 |
1146 | row-gh6h_y9cb_6knb | 00000000-0000-0000-E039-77F6B2766B82 | 0 | 1674521616 | None | 1674521616 | None | { } | 2023-01-20T00:00:00 | 0 | 22182 |
1147 | row-cb4r~rtkn.xg9x | 00000000-0000-0000-CCC4-D0C4B5D12422 | 0 | 1674521616 | None | 1674521616 | None | { } | 2023-01-21T00:00:00 | 0 | 22182 |
1148 | row-4en9-p4vi.fq5m | 00000000-0000-0000-7F40-89071F462EE8 | 0 | 1674521616 | None | 1674521616 | None | { } | 2023-01-22T00:00:00 | 0 | 22182 |
CO2 concentrations have been monitored at Mauna Loa Observatory since 1958 (website link).
co2_file = "data/co2_mm_mlo.txt"
Let's do some EDA!!
Let's instead check out this file with JupyterLab.
.txt
file.Looking at the first few lines of the data, we spot some relevant characteristics:
We can use read_csv
to read the data into a Pandas data frame, and we provide several arguments to specify that the separators are white space, there is no header (we will set our own column names), and to skip the first 72 rows of the file.
co2 = pd.read_csv(
co2_file, header = None, skiprows = 72,
sep = r'\s+' #delimiter for continuous whitespace (stay tuned for regex next lecture))
)
co2.head()
0 | 1 | 2 | 3 | 4 | 5 | 6 | |
---|---|---|---|---|---|---|---|
0 | 1958 | 3 | 1958.21 | 315.71 | 315.71 | 314.62 | -1 |
1 | 1958 | 4 | 1958.29 | 317.45 | 317.45 | 315.29 | -1 |
2 | 1958 | 5 | 1958.38 | 317.50 | 317.50 | 314.71 | -1 |
3 | 1958 | 6 | 1958.46 | -99.99 | 317.10 | 314.85 | -1 |
4 | 1958 | 7 | 1958.54 | 315.86 | 315.86 | 314.98 | -1 |
Congratulations! You've wrangled your first set of real world data!
...But our columns aren't named. We need to do more EDA.
co2.head()
0 | 1 | 2 | 3 | 4 | 5 | 6 | |
---|---|---|---|---|---|---|---|
0 | 1958 | 3 | 1958.21 | 315.71 | 315.71 | 314.62 | -1 |
1 | 1958 | 4 | 1958.29 | 317.45 | 317.45 | 315.29 | -1 |
2 | 1958 | 5 | 1958.38 | 317.50 | 317.50 | 314.71 | -1 |
3 | 1958 | 6 | 1958.46 | -99.99 | 317.10 | 314.85 | -1 |
4 | 1958 | 7 | 1958.54 | 315.86 | 315.86 | 314.98 | -1 |
co2 = pd.read_csv(
co2_file, header = None, skiprows = 72,
sep = '\s+', #regex for continuous whitespace (next lecture)
names = ['Yr', 'Mo', 'DecDate', 'Avg', 'Int', 'Trend', 'Days']
)
co2.head()
Yr | Mo | DecDate | Avg | Int | Trend | Days | |
---|---|---|---|---|---|---|---|
0 | 1958 | 3 | 1958.21 | 315.71 | 315.71 | 314.62 | -1 |
1 | 1958 | 4 | 1958.29 | 317.45 | 317.45 | 315.29 | -1 |
2 | 1958 | 5 | 1958.38 | 317.50 | 317.50 | 314.71 | -1 |
3 | 1958 | 6 | 1958.46 | -99.99 | 317.10 | 314.85 | -1 |
4 | 1958 | 7 | 1958.54 | 315.86 | 315.86 | 314.98 | -1 |
Scientific studies tend to have very clean data, right...? Let's jump right in and make a time series plot of CO2 monthly averages.
sns.lineplot(x='DecDate', y='Avg', data=co2);
The code above uses the seaborn
plotting library (abbreviated sns
).
We won't cover this library in detail until next week, so focus
on the plots themselves, not the code used to create them.
Yikes! Plotting the data uncovered a problem. It looks like we have some missing values. What happened here?
co2.head()
Yr | Mo | DecDate | Avg | Int | Trend | Days | |
---|---|---|---|---|---|---|---|
0 | 1958 | 3 | 1958.21 | 315.71 | 315.71 | 314.62 | -1 |
1 | 1958 | 4 | 1958.29 | 317.45 | 317.45 | 315.29 | -1 |
2 | 1958 | 5 | 1958.38 | 317.50 | 317.50 | 314.71 | -1 |
3 | 1958 | 6 | 1958.46 | -99.99 | 317.10 | 314.85 | -1 |
4 | 1958 | 7 | 1958.54 | 315.86 | 315.86 | 314.98 | -1 |
co2.tail()
Yr | Mo | DecDate | Avg | Int | Trend | Days | |
---|---|---|---|---|---|---|---|
733 | 2019 | 4 | 2019.29 | 413.32 | 413.32 | 410.49 | 26 |
734 | 2019 | 5 | 2019.38 | 414.66 | 414.66 | 411.20 | 28 |
735 | 2019 | 6 | 2019.46 | 413.92 | 413.92 | 411.58 | 27 |
736 | 2019 | 7 | 2019.54 | 411.77 | 411.77 | 411.43 | 23 |
737 | 2019 | 8 | 2019.62 | 409.95 | 409.95 | 411.84 | 29 |
Some data have unusual values like -1 and -99.99.
Let's check the description at the top of the file again.
Days
the equipment was in operation that month.Avg
How can we fix this? First, let's explore other aspects of our data. Understanding our data will help us decide what to do with the missing values.
First, we consider the shape of the data. How many rows should we have?
co2.shape
(738, 7)
Nice!! The number of rows (i.e. records) match our expectations.
Let's now check the quality of each feature.
Days
¶Days
is a time field, so let's analyze other time fields to see if there is an explanation for missing values of days of operation.
Let's start with months Mo
.
Are we missing any records? The number of months should have 62 or 61 instances (March 1957-August 2019).
co2["Mo"].value_counts().sort_index()
1 61 2 61 3 62 4 62 5 62 6 62 7 62 8 62 9 61 10 61 11 61 12 61 Name: Mo, dtype: int64
As expected Jan, Feb, Sep, Oct, Nov, and Dec have 61 occurrences and the rest 62.
Next let's explore days Days
itself, which is the number of days that the measurement equipment worked.
sns.displot(co2['Days']);
plt.title("Distribution of days feature")
plt.show() # suppresses unneeded plotting output
In terms of data quality, a handful of months have averages based on measurements taken on fewer than half the days. In addition, there are nearly 200 missing values--that's about 27% of the data!
Finally, let's check the last time feature, year Yr
.
Let's check to see if there is any connection between missingness and the year of the recording.
sns.scatterplot(x="Yr", y="Days", data=co2);
plt.title("Day field by Year"); # the ; suppresses output
Observations:
Potential Next Steps:
Avg
¶Next, let's return to the -99.99 values in Avg
to analyze the overall quality of the CO2 measurements.
# Histograms of average CO2 measurements
sns.displot(co2['Avg']);
The non-missing values are in the 300-400 range (a regular range of CO2 levels).
We also see that there are only a few missing Avg
values (<1% of values). Let's examine all of them:
co2[co2["Avg"] < 0]
Yr | Mo | DecDate | Avg | Int | Trend | Days | |
---|---|---|---|---|---|---|---|
3 | 1958 | 6 | 1958.46 | -99.99 | 317.10 | 314.85 | -1 |
7 | 1958 | 10 | 1958.79 | -99.99 | 312.66 | 315.61 | -1 |
71 | 1964 | 2 | 1964.12 | -99.99 | 320.07 | 319.61 | -1 |
72 | 1964 | 3 | 1964.21 | -99.99 | 320.73 | 319.55 | -1 |
73 | 1964 | 4 | 1964.29 | -99.99 | 321.77 | 319.48 | -1 |
213 | 1975 | 12 | 1975.96 | -99.99 | 330.59 | 331.60 | 0 |
313 | 1984 | 4 | 1984.29 | -99.99 | 346.84 | 344.27 | 2 |
There doesn't seem to be a pattern to these values, other than that most records also were missing Days
data.
Avg
Data?¶How should we address the invalid Avg
data?
A. Drop records
B. Set to NaN
C. Impute using some strategy
Remember we want to fix the following plot:
sns.lineplot(x='DecDate', y='Avg', data=co2)
plt.title("CO2 Average By Month");
Since we are plotting Avg
vs DecDate
, we should just focus on dealing with missing values for Avg
.
Let's consider a few options:
What do you think are the pros and cons of each possible action?
Let's examine each of these three options.
# 1. Drop missing values
co2_drop = co2[co2['Avg'] > 0]
# 2. Replace NaN with -99.99
co2_NA = co2.replace(-99.99, np.NaN)
We'll also use a third version of the data. First, we note that the dataset already comes with a substitute value for the -99.99.
From the file description:
The
interpolated
column includes average values from the preceding column (average
) and interpolated values where data are missing. Interpolated values are computed in two steps...
The Int
feature has values that exactly match those in Avg
, except when Avg
is -99.99, and then a reasonable estimate is used instead.
So, the third version of our data will use the Int
feature instead of Avg
.
# 3. Use interpolated column which estimates missing Avg values
co2_impute = co2.copy()
co2_impute['Avg'] = co2['Int']
What's a reasonable estimate?
To answer this question, let's zoom in on a short time period, say the measurements in 1958 (where we know we have two missing values).
# results of plotting data in 1958
def line_and_points(data, ax, title):
# assumes single year, hence Mo
ax.plot('Mo', 'Avg', data=data)
ax.scatter('Mo', 'Avg', data=data)
ax.set_xlim(2, 13)
ax.set_title(title)
ax.set_xticks(np.arange(3, 13))
def data_year(data, year):
return data[data["Yr"] == 1958]
# uses matplotlib subplots
# you may see more next week; focus on output for now
fig, axes = plt.subplots(ncols = 3, figsize=(12, 4), sharey=True)
year = 1958
line_and_points(data_year(co2_drop, year), axes[0], title="1. Drop Missing")
line_and_points(data_year(co2_NA, year), axes[1], title="2. Missing Set to NaN")
line_and_points(data_year(co2_impute, year), axes[2], title="3. Missing Interpolated")
fig.suptitle(f"Monthly Averages for {year}")
plt.tight_layout()
In the big picture since there are only 7 Avg
values missing (<1% of 738 months), any of these approaches would work.
However there is some appeal to option C: Imputing:
Let's replot our original figure with option 3:
sns.lineplot(x='DecDate', y='Avg', data=co2_impute)
plt.title("CO2 Average By Month, Imputed");
Looks pretty close to what we see on the NOAA website!
From the description:
The data you present depends on your research question.
How do CO2 levels vary by season?
Are CO2 levels rising over the past 50+ years, consistent with global warming predictions?
co2_year = co2_impute.groupby('Yr').mean()
sns.lineplot(x='Yr', y='Avg', data=co2_year)
plt.title("CO2 Average By Year");
Indeed, we see a rise by nearly 100 ppm of CO2 since Mauna Loa began recording in 1958.