Clone the repository and follow the directions in prepare the data. Place the dataset in the root of the repository.
Run a simple Python web server from the root of the repository:
python -m http.server 8000
Navigate to 0.0.0.0:8000 to see the choropleth.
To acquire the COVID-19 data, download the google open data. Make sure to grab the main.csv
To acquire the housing data go to Zillow and download the ZHVI by county.
This repository provides the script clean_data to create the joined dataset. It can be run with
./clean_data.py --housing <path_to_housing> --covid <path_to_covid> --output <output_file_name>
The cleaned dataset has the schema
0 date 1155225 non-null datetime64[ns]
1 county_fips 1155225 non-null object
2 population 1155225 non-null Int64
3 total_confirmed 853236 non-null Int64
4 new_confirmed 849995 non-null Int64
5 Zhvi 1043100 non-null float64
You can use your favorite method to forecast the data. We used AutoML on AWS.
./clean_data.py --housing <path_to_housing> --covid <path_to_covid> --forecast <output_file_name> --output <output_file_name>