* Corresponding author and creator
Version 1.1 | Created 2023-07-12 | Last Updated 2023-07-21
This study is a reproduction of:
Spielman, S. E., Tuccillo, J., Folch, D. C., Schweikert, A., Davies, R., Wood, N., & Tate, E. (2020). Evaluating Social Vulnerability Indicators: Criteria and their Application to the Social Vulnerability Index. Natural Hazards, 100(1), 417–436. https://doi.org/10.1007/s11069-019-03820-z
The Spielman et al. (2020) paper is in turn a replication of:
Cutter, S. L., Boruff, B. J., & Shirley, W. L. (2003). Social vulnerability to environmental hazards. Social Science Quarterly, 84(2), 242–261. https://doi.org/10.1111/1540-6237.8402002
Spielman et al. (2020) developed methods to evaluate the internal consistency and construct validity of the Cutter, Boruff and Shirley (2003) Social Vulnerability Index (SoVI). First, they reproduce a national SoVI model and validate it against an SPSS procedure provided by the original research group (Hazards Vulnerability Research Institute at University of South Carolina). The original SoVI uses 42 independent z-score normalized variables from the U.S. Census, reduces the data to factors using Principal Components Analysis, selects the first eleven factors, inverts factors with inverse relationships to social vulnerability, and sums the factors together to produce a SoVI score. The reproduced SoVI model was slightly different than the original model due to changes in U.S. Census data, using only 28 variables.
Spielman et al. modify the geographic extent of the SoVI calculation by calculating SoVI on a national extent, and then recalculating for each of ten Federal Emergency Management Agency (FEMA) regions, and again for a single state or cluster of states within each of the ten regions, resulting in 21 total indices. Internal consistency is assessed by calculating the spearman rank correlation coefficient of the SoVI score for counties in the state model compared to the FEMA region model and national model. Construct validity is assessed by summing the loadings for each input variable across the PCA factors in each model and calculating the variables sign (positive/negative) and the rank of the variable's total loading compared to the other variables. These signs and ranks are summarized across all 21 versions of the SoVI model with regard to the number of times the sign is different from the national model and the distributions of ranks.
In this reproduction study, we attempt to reproduce identical SoVI model outputs for each of the 21 models in the original study. We will compare these outputs to data files in Spielman et al.'s GitHub repository. We will also attempt to reproduce identical results of internal consistency analysis (figure 1 and table 2) and construct validity analysis (figure 2) from Spielman et al.'s paper. We succeed in reproducing identical SoVI model outputs, but find slight discrepancies in our figures and tables.
The code in this Jupyter notebook report is adapted from Spielman et al.'s GitHub repository. The original study states the intended open source permissions in the acknowledgements: "To facilitate advances to current practice and to allow replication of our results, all of the code and data used in this analysis is open source and available at (https://github.com/geoss/sovi-validity). Funding was provided by the US National Science Foundation (Award No. 1333271) and the U.S. Geological Survey Land Change Science Program."
Social vulnerability, social indicators, Principal Component Analysis, reproducibility
We computationally reproduce Spielman et al.'s original work using the code provided in their Github repository (https://github.com/geoss/sovi-validity), adapting their code to run in an updated Python environment using current package versions. We make all of our work available online using the HEGSRR reproducible research compendium template.
The original paper was a replication study testing the sensitivity of SoVI to changes in geographic extent. Spielman et al. addressed the following hypotheses in their work:
OR-H1: SoVI is internally inconsistent.
To address this hypothesis, Spielman et al. illustrated that SoVI is not robust to changes in geographic extent by calculating SoVI scores for ten selected states or groups of states on three geographic extents: national, FEMA region, and state(s). The counties within the state(s) of interest were then selected and ranked according to their SoVI score. OR-H1 was tested by calculating Spearman's rank correlation between the state and FEMA region models and between the state and national models.
OR-H2: SoVI is theoretically inconsistent.
To address this hypothesis, Spielman et al. used the same SoVI models as described under OR-H1. For each model, they summed all of the PCA factors together to determine the net influence of each variable in each model. Then they recorded the signs of each variable and calculated the number of deviations of the ten state and FEMA region models from the national model. They also ranked the variables by absolute value for each model and calculated summary statistics regarding the distribution of ranks for each variable amongst all models. Spielman et al. did not use a particular statistical method to test OR-H2, but illustrated substantial disagreements between variable rankings and signs amongst the 21 SoVI models.
For our reproduction, we address the following three hypotheses:
RPr-H1: Reproduced SoVI model scores and other reproduced output datasets are not identical to the original study SoVI model scores and provided output datasets for each of the 21 SoVI models.
RPr-H2: Reproduced figures and tables for the internal consistency analysis are not identical to the figures and tables (figure 1 and table 2) of the original study.
RPr-H3: For the theoretical consistency analysis, reproduced direction reversals and min, average, and max SoVI rank value of 28 demographic variables are not identical to the direction reversals and min, average, and max SoVI rank values shown in figure 2 of the original study.
We answer these questions by working through Spielman et al.'s code line by line in an updated python coding environment. To improve reproducibility, we reorganize Spielman's repository into the Template for Reproducible and Replicable Research in Human-Environment and Geographical Sciences (doi:10.17605/OSF.IO/W29MQ) and use one Jupyter notebook for the reproduction report and code. We catalogue barriers to reproducibility and make improvements wherever possible.
Disclaimer: we worked with the data and code before writing this report, so there is no pre-registration of the analysis plan. We originally intended to publish only a replication of this study; we did not anticipate publishing a reproduction until we spent some time working with the code.
We made several modifications to this reproduction during our analysis. First, we visualized the nationwide SoVI model using an interactive leaflet map. This allowed us to see the general spatial trends of the social vulnerability model across the US. A more significant modification that we tested was calculating the percentage of variance in SoVI explained by each component of the model. This calculation was not done by in our original reproduction and allowed us to get a sense of which factors in the model were having the largest impact on the model itself. We then weighted the model factors by their percent variance explained such that the factors that explained the least variance had very little impact on the final model output and the factors that explained the most variance had the largest impact on the final model.
Spatial Coverage
: United States, excluding Puerto RicoSpatial Resolution
: Counties and county equivalentsSpatial Reference System
: EPSG:4269Temporal Coverage
: 2008 - 2012 (data is the 2012 5-year ACS)Temporal Resolution
: One-time measurement, does not address change over timeCurrently, we are using a 2020 MacBook Pro running on macOS Ventura 13.3.1. We anticipate collaborators working on the project from different computers and different operating systems, and we seek to containerize the project so that scripts can be run on many different machines.
The original study used Python for their analysis, so we reproduce their results in Python, using a containerized conda environment. This environment consists of Python 3.9.16 and the software packages listed in requirements.txt
To set up this environment on another machine, one should install the correct version of Python and run the cell below to install the correct package versions. If a user wishes to create a self-contained environment, they should explore venv, conda, or pipenv virtual environments.
# report python version and install required packages
# switch if statement from True to False once packages have been installed
if True:
!python -V
!pip install -r ../environment/requirements.txt
# Import modules, define directories
import pygris
import pandas as pd
import geopandas as gpd
from pygris.data import get_census
from pygris import counties
from pyhere import here
import numpy as np
import libpysal as lps
import lxml
import tabulate
from scipy.stats import spearmanr
from scipy.stats.mstats import zscore as ZSCORE
from scipy.stats import rankdata
import mdp as MDP
from operator import itemgetter
import copy
from matplotlib.colors import ListedColormap
from matplotlib import patheffects as pe
import matplotlib.pyplot as plt
from IPython import display
from IPython.display import Markdown, Latex
import folium
pd.set_option("chained_assignment", None)
path = {
"dscr": here("data", "scratch"),
"drpub": here("data", "raw", "public", "spielman", "input"),
"drpub2": here("data", "raw", "public"),
"drpriv": here("data", "raw", "private"),
"ddpub": here("data", "derived", "public", "version1"),
"ddpriv": here("data", "derived", "private"),
"rfig": here("results", "figures"),
"roth": here("results", "other"),
"rtab": here("results", "tables"),
"og_out": here("data", "raw", "public", "spielman", "output"),
"dmet": here("data", "metadata")
}
# Switch from False to True to regenerate documentation of computational environment
# Note that this approach is not perfect -- it may miss some packages
# This code may work better from the command prompt
if False:
!pip install pigar
!pigar generate -f ../environment/requirements.txt
For Spielman et al.'s original study, the data sources were the 2008-2012 5-year American Community Survey and the 2010 decennial census. Spielman et al. downloaded their data from Social Explorer; in our reproduction, we pull our data directly from the census into Python via a census API package known as pygris. These variables are based on the original work by Cutter et al. to create SoVI, and cover a wide range of social and demographic information, the particulars of which are described below.
In order to confirm that our data and Spielman et al.'s data perfectly match each other, we import the names of relevant variables from both datasets here.
# Import data dictionary
acs_vars = pd.read_csv( here("data", "metadata", "ACS_2012_data_dictionary.csv") )
acs_vars.drop(columns=acs_vars.columns[0], axis=1, inplace=True)
acs_variables = list(acs_vars['Reproduction Label'][1:])
spielman_acs_variables = list(acs_vars['Spielman Label'][1:])
Used in both original study and reproduction.
Planned deviation: to enhance reproducibility, we draw the data directly from the census into python using the pygris package.
# Switch from False to True to download fresh data from the Census
if False:
# Acquire attribute data for reproduction
counties_detailed = get_census(dataset = "acs/acs5", # dataset name on the Census API you are connecting to; find datasets at https://api.census.gov/data.html
variables = acs_variables, # string (or list of strings) of desired vars. For the 2021 5-year ACS Data Profile, those variable IDs are found at https://api.census.gov/data/2021/acs/acs5/profile/variables.html
year = 2012, # year of your data (or end-year for a 5-year ACS sample)
params = { # dict of query parameters to send to the API.
"for": "county:*"},
guess_dtypes = True,
return_geoid = True)
# Drop Puerto Rico
counties_detailed = counties_detailed.loc[~counties_detailed['GEOID'].str.startswith('72')]
# Download and save raw data
counties_detailed.to_csv( here(path["drpub2"], "counties_attributes_raw.csv"))
else:
counties_detailed = pd.read_csv( here(path["drpub2"], "counties_attributes_raw.csv"), dtype = {'GEOID': object} )
counties_detailed = counties_detailed.drop(counties_detailed.columns[0],axis=1)
Load data from Spielman et al.'s research repository for validation of the reproduction study.
# Import original data from Spielman et al.
# Import base ACS data
make_strings = {'Geo_FIPS': object, 'Geo_STATE': object, 'Geo_COUNTY': object,
'Geo_TRACT': object, 'Geo_CBSA': object, 'Geo_CSA': object}
acs = pd.read_csv(here(path["drpub"], 'sovi_acs.csv'),
dtype=make_strings, skiprows=1,encoding='latin-1')
# Import, join an ACS supplemental
acs_sup2 = pd.read_csv(here(path["drpub"], 'sovi_acs_kids.csv'),
dtype=make_strings, skiprows=1,encoding='latin-1')
acs = acs.merge(acs_sup2, how = "inner", on='Geo_FIPS')
# Drop Puerto Rico
acs = acs[acs.Geo_STATE_x != '72']
Markdown( here(path["dmet"], "ACS_2012_geographic_metadata.md") )
Title
: American Community Survey 2012 5-year Estimate Demographic VariablesAbstract
: The 5-year ACS provides estimates surrounding demographic information in the USA. These estimates are more reliable than 1-year and 3-year estimates but less reliable than decennial census data. On the other hand, 5-year estimates are less current than 1-year and 3-year estimates because they represent measurements taken over 60 months. See the census website for more details.Spatial Coverage
: United States, excluding Puerto RicoSpatial Resolution
: County and county-equivalentsSpatial Reference System
: None, just attribute dataTemporal Coverage
: 2008-2012Temporal Resolution
: Data averaged over five yearsLineage
: Original data downloaded from Social Explorer and then placed in the original study's GitHub repository: https://github.com/geoss/sovi-validity. Reproduction data obtained directly from the census via API.Distribution
: The reproduction data is distributed via a census API. See the detailed tables on the census website and instructions for drawing census data directly into python on the pygris website. Spielman et al. originally accessed the ACS data with Social Explorer from the following two tables.
Constraints
: Census data is available in the public domainData Quality
: Margin of error provided by the Census Bureau for relevant variablesVariables
: See ACS_2012_data_dictionary.csvacs_vars
Reproduction Label | Spielman Label | Alias | Definition | Type | Domain | Missing Data Value(s) | Missing Data Frequency | |
---|---|---|---|---|---|---|---|---|
0 | GEOID | Geo_FIPS | FIPS code unique identifier | Unique code for every county and county-equiva... | string | 01001 - 56045 | None | 0.0 |
1 | B01002_001E | ACS12_5yr_B01002001 | median age | MEDIAN AGE BY SEX: Estimate!!Median age!!Total | float64 | 21.7 - 63 | NaN | 0.0 |
2 | B03002_001E | ACS12_5yr_B03002001 | total population of respondents to race/ethnicity | HISPANIC OR LATINO ORIGIN BY RACE: Estimate!!T... | int64 | 66 - 9840024 | NaN | 0.0 |
3 | B03002_004E | ACS12_5yr_B03002004 | total Black population | HISPANIC OR LATINO ORIGIN BY RACE: Estimate!!T... | int64 | 0 - 1267825 | NaN | 0.0 |
4 | B03002_005E | ACS12_5yr_B03002005 | total Native American population | HISPANIC OR LATINO ORIGIN BY RACE: Estimate!!T... | int64 | 0 - 59060 | NaN | 0.0 |
5 | B03002_006E | ACS12_5yr_B03002006 | total Asian population | HISPANIC OR LATINO ORIGIN BY RACE: Estimate!!T... | int64 | 0 - 1343920 | NaN | 0.0 |
6 | B03002_012E | ACS12_5yr_B03002012 | total Latinx population | HISPANIC OR LATINO ORIGIN BY RACE: Estimate!!T... | int64 | 0 - 4694846 | NaN | 0.0 |
7 | B06001_002E | ACS12_5yr_B06001002 | total population under 5 years of age | PLACE OF BIRTH BY AGE IN THE UNITED STATES: Es... | float64 | 0 - 651662 | NaN | 78.0 |
8 | B09020_001E | ACS12_5yr_B09020001 | total population over 65 years of age | RELATIONSHIP BY HOUSEHOLD TYPE (INCLUDING LIVI... | int64 | 5 - 1078555 | NaN | 0.0 |
9 | B01003_001E | ACS12_5yr_B01003001 | total population | TOTAL POPULATION: Estimate!!Total | int64 | 66 - 9840024 | NaN | 0.0 |
10 | B25008_001E | ACS12_5yr_B25008001 | total population in occupied housing units | TOTAL POPULATION IN OCCUPIED HOUSING UNITS BY ... | int64 | 62 - 9664175 | NaN | 0.0 |
11 | B25002_002E | ACS12_5yr_B25002002 | total occupied housing units | OCCUPANCY STATUS: Estimate!!Total!!Occupied | int64 | 35 - 3218511 | NaN | 0.0 |
12 | B25003_003E | ACS12_5yr_B25003003 | total renter occupied housing units | TENURE: Estimate!!Total!!Renter occupied | int64 | 14 - 1695180 | NaN | 0.0 |
13 | B25002_001E | ACS12_5yr_B25002001 | total housing units for which occupancy status... | OCCUPANCY STATUS: Estimate!!Total | int64 | 70 - 3441416 | NaN | 0.0 |
14 | B09020_021E | ACS12_5yr_B09020021 | total 65+ living in group quarters | RELATIONSHIP BY HOUSEHOLD TYPE (INCLUDING LIVI... | int64 | 0 - 37611 | NaN | 0.0 |
15 | B01001_026E | ACS12_5yr_B01001026 | total female population | SEX BY AGE: Estimate!!Total!!Female | int64 | 20 - 4987765 | NaN | 0.0 |
16 | B11001_006E | ACS12_5yr_B11001006 | total female-headed family households | HOUSEHOLD TYPE (INCLUDING LIVING ALONE): Estim... | int64 | 0 - 498851 | NaN | 0.0 |
17 | B11001_001E | ACS12_5yr_B11001001 | total households for which household type is k... | HOUSEHOLD TYPE (INCLUDING LIVING ALONE): Estim... | int64 | 35 - 3218511 | NaN | 0.0 |
18 | B25002_003E | ACS12_5yr_B25002003 | total vacant housing units | OCCUPANCY STATUS: Estimate!!Total!!Vacant | int64 | 35 - 245069 | NaN | 0.0 |
19 | B19025_001E | ACS12_5yr_B19025001 | aggregate household income | AGGREGATE HOUSEHOLD INCOME IN THE PAST 12 MONT... | int64 | 1785600 - 263044380000 | NaN | 0.0 |
20 | B23022_025E | ACS12_5yr_B23022025 | total males unemployed for last 12 months | SEX BY WORK STATUS IN THE PAST 12 MONTHS BY US... | int64 | 1 - 726803 | NaN | 0.0 |
21 | B23022_049E | ACS12_5yr_B23022049 | total females unemployed for last 12 months | SEX BY WORK STATUS IN THE PAST 12 MONTHS BY US... | int64 | 0 - 1131737 | NaN | 0.0 |
22 | B23022_001E | ACS12_5yr_B23022001 | total population for which unemployment and se... | SEX BY WORK STATUS IN THE PAST 12 MONTHS BY US... | int64 | 45 - 6658456 | NaN | 0.0 |
23 | B17021_002E | ACS12_5yr_B17021002 | total population below poverty level | POVERTY STATUS OF INDIVIDUALS IN THE PAST 12 M... | int64 | 0 - 1658231 | NaN | 0.0 |
24 | B17021_001E | ACS12_5yr_B17021001 | total population for which poverty information... | POVERTY STATUS OF INDIVIDUALS IN THE PAST 12 M... | int64 | 64 - 9684503 | NaN | 0.0 |
25 | B25024_010E | ACS12_5yr_B25024010 | number of mobile home housing units in structure | UNITS IN STRUCTURE: Estimate!!Total!!Mobile home | int64 | 0 - 85377 | NaN | 0.0 |
26 | B25024_001E | ACS12_5yr_B25024001 | total housing units in structure | UNITS IN STRUCTURE: Estimate!!Total | int64 | 70 - 3441416 | NaN | 0.0 |
27 | C24010_038E | ACS12_5yr_C24010038 | total female employed | SEX BY OCCUPATION FOR THE CIVILIAN EMPLOYED PO... | int64 | 12 - 2056023 | NaN | 0.0 |
28 | C24010_001E | ACS12_5yr_C24010001 | total population for which sex and occupation ... | SEX BY OCCUPATION FOR THE CIVILIAN EMPLOYED PO... | int64 | 54 - 4495118 | NaN | 0.0 |
29 | B19055_002E | ACS12_5yr_B19055002 | total households with social security income | SOCIAL SECURITY INCOME IN THE PAST 12 MONTHS F... | int64 | 9 - 726298 | NaN | 0.0 |
30 | B19055_001E | ACS12_5yr_B19055001 | total households for which social security inc... | SOCIAL SECURITY INCOME IN THE PAST 12 MONTHS F... | int64 | 35 - 3218511 | NaN | 0.0 |
31 | B09002_002E | ACS12_5yr_B09002002 | total children in married couple families | OWN CHILDREN UNDER 18 YEARS BY FAMILY TYPE AND... | int64 | 0 - 1380977 | NaN | 0.0 |
32 | B09002_001E | ACS12_5yr_B09002001 | total children for which family type and age a... | OWN CHILDREN UNDER 18 YEARS BY FAMILY TYPE AND... | int64 | 0 - 2032147 | NaN | 0.0 |
33 | B19001_017E | ACS12_5yr_B19001017 | total households with over 200k income | HOUSEHOLD INCOME IN THE PAST 12 MONTHS (IN 201... | int64 | 0 - 208954 | NaN | 0.0 |
34 | B06007_005E | ACS12_5yr_B06007005 | total Spanish-speakers who speak english less ... | PLACE OF BIRTH BY LANGUAGE SPOKEN AT HOME AND ... | float64 | 0 - 1695391 | NaN | 78.0 |
35 | B06007_008E | ACS12_5yr_B06007008 | total people who speak another language and sp... | PLACE OF BIRTH BY LANGUAGE SPOKEN AT HOME AND ... | float64 | 0 - 743418 | NaN | 78.0 |
36 | B06007_001E | ACS12_5yr_B06007001 | total population with known language spoken at... | PLACE OF BIRTH BY LANGUAGE SPOKEN AT HOME AND ... | float64 | 66 - 9188362 | NaN | 78.0 |
37 | B16010_002E | ACS12_5yr_B16010002 | total population with less than a high school ... | EDUCATIONAL ATTAINMENT AND EMPLOYMENT STATUS B... | int64 | 5 - 1508273 | NaN | 0.0 |
38 | B16010_001E | ACS12_5yr_B16010001 | total for which education, employment, languag... | EDUCATIONAL ATTAINMENT AND EMPLOYMENT STATUS B... | int64 | 65 - 6380366 | NaN | 0.0 |
39 | C24050_002E | ACS12_5yr_C24050002 | total population in extractive industries | INDUSTRY BY OCCUPATION FOR THE CIVILIAN EMPLOY... | int64 | 0 - 54942 | NaN | 0.0 |
40 | C24050_001E | ACS12_5yr_C24050001 | total population for which industry known | INDUSTRY BY OCCUPATION FOR THE CIVILIAN EMPLOY... | int64 | 54 - 4495118 | NaN | 0.0 |
41 | C24050_029E | ACS12_5yr_C24050029 | total people in service occupations | INDUSTRY BY OCCUPATION FOR THE CIVILIAN EMPLOY... | int64 | 4 - 837607 | NaN | 0.0 |
42 | B08201_002E | ACS12_5yr_B08201002 | total households with no available vehicle | HOUSEHOLD SIZE BY VEHICLES AVAILABLE: Estimate... | int64 | 0 - 577967 | NaN | 0.0 |
43 | B08201_001E | ACS12_5yr_B08201001 | total households for which vehicle status and ... | HOUSEHOLD SIZE BY VEHICLES AVAILABLE: Estimate... | int64 | 35 - 3218511 | NaN | 0.0 |
44 | B25064_001E | ACS12_5yr_B25064001 | median gross rent | MEDIAN GROSS RENT (DOLLARS): Estimate!!Median ... | int64 | 275 - 1678 | NaN | 0.0 |
45 | B25077_001E | ACS12_5yr_B25077001 | median home value | MEDIAN VALUE (DOLLARS): Estimate!!Median value... | float64 | 19400 - 944100 | NaN | 1.0 |
The above are the metadata files that we wrote for our pygris-acquired version of this data, stored as ACS_2012_data_dictionary.csv and ACS_2012_geographic_metadata.md. The metadata files provided by Spielman et al. are also in our repository, named sovi_acs.txt and sovi_acs_kids.txt.
Used in Spielman et al.'s original study.
# Import decennial supplemental
dec_sup1 = pd.read_csv(here(path["drpub"],'sovi_decennial_sup1.csv'),
dtype=make_strings,skiprows=1,encoding='latin-1')
Markdown( here(path["dmet"], "dec_2010_metadata.md") )
Title
: 2010 Decennial CensusAbstract
: Collected once every ten years, the decennial census documents demographic and population data in the United States.Spatial Coverage
: United States, excluding Puerto RicoSpatial Resolution
: County and county-equivalentsSpatial Reference System
: None, just attribute dataTemporal Coverage
: 2010Temporal Resolution
: One-time observationsLineage
: Original data downloaded from Social Explorer and then placed in the original study's GitHub repository: https://github.com/geoss/sovi-validity.Distribution
: Visit this URL for accessConstraints
: Census data is available in the public domainData Quality
: Margin of error provided by the Census Bureau for relevant variablesVariables
:Label | Alias | Definition | Type | Accuracy | Domain | Missing Data Value(s) | Missing Data Frequency |
---|---|---|---|---|---|---|---|
SE_T02A_002 | Land area | Area (Land) in square miles | float64 | ... | 1.998779 - 145504.8 | nan | 0 |
Geo_FIPS | FIPS code unique identifier | Unique code for every county and county-equivalent in USA | string | ... | g01001 - g56045 | None | 0 |
Original metadata file provided by Spielman et al. as sovi_decennial_sup1.txt.
Used in Spielman et al.'s original study
spielman_geom = gpd.read_file( here(path["drpub"], "USA_Counties_500k.shp") )
Markdown( here(path["dmet"], 'USA_counties_metadata.md') )
Title
: USA Counties Geographic ShapefileAbstract
: No metadata provided, so it is unclear exactly where Spielman et al. acquired this file but they likely downloaded it directly or indirectly from the census. The shapefile provides the geometries of counties and county-equivalents in the United States, with limited attribute information including county name and a unique identifier.Spatial Coverage
: United States, excluding Puerto RicoSpatial Resolution
: County and county-equivalentsSpatial Reference System
: EPSG:4269Temporal Coverage
: UnknownTemporal Resolution
: One-time observationsLineage
: UnknownDistribution
: Unknown. Presumably downloaded from census.Constraints
: Census data is available in the public domainData Quality
: 1:500,000 scaleVariables
: For each variable, enter the following information. If you have two or more variables per data source, you may want to present this information in table form (shown below)Label | Alias | Definition | Type | Accuracy | Domain | Missing Data Value(s) | Missing Data Frequency |
---|---|---|---|---|---|---|---|
geoFIPS | FIPS code unique identifer | Unique code for every county and county-equivalent in USA | string | ... | g01001 - g56045 | None | 0 |
No original metadata file provided.
Used in reproduction study.
Planned deviation: to enhance reproducibility, we draw the data directly from the census into python using the pygris package.
# Switch False to True if you wish to acquire data directly from census
if False:
# Acquire geographical data for reproduction
counties_shp = counties(cb = True, year = 2010, cache = True) # year 2012 (and 2011) cartographic boundaries not available
# Save raw data
counties_shp.to_file( here(path["drpub2"], "counties_geometries_raw.gpkg") )
else:
counties_shp = gpd.read_file( here(path["drpub2"], "counties_geometries_raw.gpkg") )
Markdown( here(path["dmet"], "county_geom_2010_metadata.md") )
Title
: USA Counties Cartographic BoundariesAbstract
: The cartographic boundary files provided by the US census are simplified representations of the MAF/TIGER files. We use the 2010 boundary file because the census has not made such a file available for 2012 or 2011 and the original paper also used land area from 2010. This shapefile provides the geometries of counties and county-equivalents in the United States, with limited attribute information including land area.Spatial Coverage
: United States, excluding Puerto RicoSpatial Resolution
: County and county-equivalentsSpatial Reference System
: EPSG:4269Temporal Coverage
: 2010Temporal Resolution
: One-time observationsLineage
: We use pygris to pull the data directly from the census into python.Distribution
: This file is distributed via a census API. See more information on the census website and instructions for drawing census data directly into python on the pygris website.Constraints
: Census data is available in the public domainData Quality
: 1:500,000 scaleVariables
:Label | Alias | Definition | Type | Domain | Missing Data Value(s) | Missing Data Frequency |
---|---|---|---|---|---|---|
STATE | State-level FIPS code | State-level FIPS code | string | 01 - 56 | None | 0 |
COUNTY | County-level FIPS code | County-level FIPS code | string | 001 - 840 | None | 0 |
CENSUSAREA | land area | land area in square miles | float64 | 1.999 - 145504.789 | NaN | 0 |
The metadata file for this data is stored as county_geom_2010_metadata.md.
A workflow diagram for this section is displayed below.
We begin with step P1: joining the geometry and attribute data
# Step P1
# Join geometry and attribute data for reproduction
counties_shp['GEOID'] = counties_shp.STATE + counties_shp.COUNTY
counties = counties_shp.merge(counties_detailed, how = "inner", on = "GEOID")
# Also join Spielman's land area information to the rest of Spielman's data
# (to check that all data is accurate, not for purposes of analysis)
acs = acs.merge(dec_sup1, how = "inner", on='Geo_FIPS')
Planned deviation: Because we decided to acquire our data independently from Spielman et al., we need to check that our data is indeed the same as theirs.
To begin, we define a function that can check that the entries of two pandas DataFrames are equal.
# Define a function that can determine whether every entry in specified columns of two tables match
def equiv(table1, sort1, column1, table2, sort2, column2):
'''
Tests two tables to see whether corresponding columns have equivalent entries.
Parameters:
table1 - the first table
sort1 - the column in the first table to join by (str)
column1 - the column(s) in the first table to test the values of (list of str) (should list analogous columns for columns2)
table2 - the second table
sort2 - the column in the second table to join by (str)
column2 - the column(s) in the second table to test the values of (list of str)
'''
# Sort tables
table1 = table1.sort_values(by = sort1).reset_index()
table2 = table2.sort_values(by = sort2).reset_index()
# Rename column name in table2 to match that in table1
for i in range(len(column1)):
table2 = table2.rename(columns={column2[i]: column1[i]})
# Select the columns to test equivalency of
table1 = table1[column1]
table2 = table2[column1]
# Perform equivalency test
test = table1.eq(table2)
return test
Next, we round our area columns to the nearest integer, just for the purposes of comparing the two columns. These columns came from different sources and we know that they do not match up exactly.
# # Round area column
# acs['SE_T02A_002_check'] = acs.SE_T02A_002.round(0)
# counties['CENSUSAREA_check'] = counties.CENSUSAREA.round(0)
# Add the area variables to the lists of variables
acs_variables.append('CENSUSAREA')
spielman_acs_variables.append('SE_T02A_002')
# Perform equivalency test
test = equiv(counties, "GEOID", acs_variables, acs, "Geo_FIPS", spielman_acs_variables)
matching_cols = pd.DataFrame({"test": test.sum().eq(3143)}) # 3143 matches the number of rows
matching_cols.loc[~matching_cols.test] # Identify the columns that have some data discrepencies
test | |
---|---|
B25077_001E | False |
CENSUSAREA | False |
The following variables have some discrepancy between the original and reproduction data:
# Inspect the data values for B25077_001E at the indices with data discrepancies
# Find the rows for which there are data discrepancies
messed_up_indices = test[["B25077_001E"]].loc[~test.B25077_001E]
# Select the data of interest from tygris
tygris_data = counties.sort_values(by = "GEOID")\
.reset_index().loc[messed_up_indices.index]\
[["GEOID", "B25077_001E"]]
# Select the data of interest from Spielman et al.
spielman_data = acs.sort_values(by = "Geo_FIPS")\
.reset_index().loc[messed_up_indices.index]\
[["Geo_FIPS", "ACS12_5yr_B25077001"]]
# Join and inspect
merged = tygris_data.merge(spielman_data, how = "inner", left_on = "GEOID", right_on = "Geo_FIPS")
merged
GEOID | B25077_001E | Geo_FIPS | ACS12_5yr_B25077001 | |
---|---|---|---|---|
0 | 15005 | NaN | 15005 | NaN |
By inspection, we see that the one disagreement between B25077_001E and ACS12_5yr_B25077001 occurs because of a NaN value in an analogous location in each of the two datasets. Rather than revealing an issue in matching our data, this shows us that we will need to impute a missing value for one NaN in B25077_001E -- median home value (see P3).
# Inspect the data values for CENSUSAREA at the indices with data discrepancies
# Find the rows for which there are data discrepancies
messed_up_indices = test[["CENSUSAREA"]].loc[~test.CENSUSAREA]
# Select the data of interest from tygris
tygris_data = counties.sort_values(by = "GEOID")\
.reset_index().loc[messed_up_indices.index]\
[["GEOID", "CENSUSAREA"]]
# Select the data of interest from Spielman et al.
spielman_data = acs.sort_values(by = "Geo_FIPS")\
.reset_index().loc[messed_up_indices.index]\
[["Geo_FIPS", "SE_T02A_002"]]
# Join and inspect
merged = tygris_data.merge(spielman_data, how = "inner", left_on = "GEOID", right_on = "Geo_FIPS")
merged
GEOID | CENSUSAREA | Geo_FIPS | SE_T02A_002 | |
---|---|---|---|---|
0 | 01001 | 594.436 | 01001 | 594.4361 |
1 | 01005 | 884.876 | 01005 | 884.8763 |
2 | 01007 | 622.582 | 01007 | 622.5823 |
3 | 01009 | 644.776 | 01009 | 644.7759 |
4 | 01011 | 622.805 | 01011 | 622.8047 |
... | ... | ... | ... | ... |
2285 | 55137 | 626.153 | 55137 | 626.1533 |
2286 | 55141 | 793.116 | 55141 | 793.1165 |
2287 | 56013 | 9183.814 | 56013 | 9183.8130 |
2288 | 56023 | 4076.129 | 56023 | 4076.1300 |
2289 | 56037 | 10426.649 | 56037 | 10426.6500 |
2290 rows × 4 columns
There are many disagreements between CENSUSAREA and SE_T02A_002, but they appear to be relatively minor differences. Let us evaluate how large those differences are.
merged["Difference"] = abs(merged["CENSUSAREA"] - merged["SE_T02A_002"])
print("Maximum difference:", merged['Difference'].max(), "\nAverage difference:", merged['Difference'].mean())
Maximum difference: 0.010999999998603016 Average difference: 0.00031027379912322995
The largest discrepency between the two different sources of land area data is just over 0.01 square miles and the average difference (amongst those with a difference) is tiny. With such a minor difference between our data and theirs, for our purposes we will consider our data roughly the same as Spielman et al.'s.
# Step P2
# Calculating the variables used in SoVI
counties['MEDAGE_ACS'] = counties.B01002_001E
counties['BLACK_ACS'] = counties.B03002_004E / (counties.B03002_001E)
counties['QNATAM_ACS'] = counties.B03002_005E / (counties.B03002_001E)
counties['QASIAN_ACS'] = counties.B03002_006E / (counties.B03002_001E)
counties['QHISP_ACS'] = counties.B03002_012E / (counties.B03002_001E)
counties['QAGEDEP_ACS'] = (counties.B06001_002E + counties.B09020_001E) / (counties.B01003_001E)
counties['QPUNIT_ACS'] = counties.B25008_001E / (counties.B25002_002E)
counties['PRENTER_ACS'] = counties.B25003_003E / (counties.B25002_001E)
counties['QNRRES_ACS'] = counties.B09020_021E / (counties.B01003_001E)
counties['QFEMALE_ACS'] = counties.B01001_026E / (counties.B01003_001E)
counties['QFHH_ACS'] = counties.B11001_006E / (counties.B11001_001E)
counties['QUNOCCHU_ACS'] = counties.B25002_003E / (counties.B25002_001E)
counties['QCVLUN'] = (counties.B23022_025E + counties.B23022_049E) / \
counties.B23022_001E
counties['QPOVTY'] = (counties.B17021_002E) / counties.B17021_001E
counties['QMOHO'] = (counties.B25024_010E) / counties.B25024_001E
counties['QFEMLBR'] = (counties.C24010_038E) / counties.C24010_001E
counties['QSSBEN'] = (counties.B19055_002E) / counties.B19055_001E
counties['QFAM'] = (counties.B09002_002E) / counties.B09002_001E
counties['QRICH200K'] = (counties.B19001_017E) / counties.B11001_001E
counties['PERCAP_ALT'] = counties.B19025_001E / (counties.B25008_001E)
counties['QESL_ALT'] = (counties.B06007_005E + counties.B06007_008E) / \
counties.B06007_001E
counties['QED12LES_ALT'] = (counties.B16010_002E) / counties.B16010_001E
counties['QEXTRCT_ALT'] = (counties.C24050_002E) / counties.C24050_001E
counties['QSERV_ALT'] = (counties.C24050_029E) / counties.C24050_001E
counties['QNOAUTO_ALT'] = (counties.B08201_002E) / counties.B08201_001E
counties['MDGRENT_ALT'] = counties.B25064_001E
counties['MHSEVAL_ALT'] = counties.B25077_001E
counties['POPDENS'] = counties.B01003_001E / (counties.CENSUSAREA)
As noted before, B25077_001E is missing a data value. We now perform one final check to see if we need to impute anything else.
# Check for missing data
for i in counties.columns:
x = counties[i].isnull().sum()
if x > 0:
print(i, "contains", x, "missing value(s).")
# Check for infinities
counties_num = counties.select_dtypes(include=['int64','float64'])
for i in counties_num.columns:
xmin = counties_num[i].min()
xmax = counties_num[i].max()
if xmin == -np.inf:
print(i, "contains a negative infinity")
elif xmax == np.inf:
print(i, "contains a positive infinity")
LSAD contains 2 missing value(s). B25077_001E contains 1 missing value(s). QFAM contains 2 missing value(s). MHSEVAL_ALT contains 1 missing value(s).
There are four variables with missing data. LSAD is not used in our analysis, so we may ignore this. B25077_001E and MHSEVAL_ALT are literally identical, so we will ignore B25077_001E and simply impute for MHSEVAL_ALT's one missing value. We also need to impute for QFAM's 2 missing values. We use the same imputation decisions that Spielman et al. employ in their analysis.
Unplanned deviation: When imputing for MHSEVAL_ALT's missing data, we removed a fair amount of extraneous code that was filling in missing spatial lag data with original data for the county. This was unnecessary because we only needed to impute data for one county and that county had spatial lag data. Also note that Spielman et al.'s method for imputing data for MHSEVAL_ALT is a deviation from Cutter's original methodology, in which she imputed a 0 for any missing value. While this is a deviation from the original SoVI methodology, 0 is an unrealistic median home value, so Spielman et al.'s method for imputation seems like a reasonable improvement.
# Step P3
# Replace missing QFAM data with 0
counties.QFAM = counties.QFAM.replace([np.inf, -np.inf, np.nan], 0)
# Replace missing MHSEVAL_ALT data with its spatial lag
# Calculate spatial weights matrix
w = lps.weights.Queen.from_dataframe(counties)
w.transform = 'R'
# Calculate spatial lag
counties['MHSEVAL_ALT_LAG'] = lps.weights.lag_spatial(w, counties.MHSEVAL_ALT)
# Impute for the missing value
counties.MHSEVAL_ALT[np.isnan(counties['MHSEVAL_ALT'])] = counties[["MHSEVAL_ALT_LAG"]][pd.isna(counties['MHSEVAL_ALT'])]
('WARNING: ', 68, ' is an island (no neighbors)') ('WARNING: ', 546, ' is an island (no neighbors)') ('WARNING: ', 547, ' is an island (no neighbors)') ('WARNING: ', 549, ' is an island (no neighbors)') ('WARNING: ', 1226, ' is an island (no neighbors)') ('WARNING: ', 1876, ' is an island (no neighbors)') ('WARNING: ', 2976, ' is an island (no neighbors)')
/Users/colmanbashore/anaconda3/envs/spielman/lib/python3.9/site-packages/libpysal/weights/weights.py:172: UserWarning: The weights matrix is not fully connected: There are 10 disconnected components. There are 7 islands with ids: 68, 546, 547, 549, 1226, 1876, 2976. warnings.warn(message)
The missing data procedure attempts to fill some counties' missing data with the average values of the surrounding counties. This procedure produces a warning about 10 disconnected components and 7 islands in the weights matrix. The cause of this error is areas of the United States that are not contiguous with one another. Fortunately, these areas are not missing any data, and therefore this error does not affect our procedure to fill gaps in missing data.
Before adjusting directionality, let us check that all of our derived variables match all of Spielman et al.'s derived variables.
# Import Spielman et al.'s derived variables
US_All = pd.read_csv(here("data", "raw", "public", "spielman", "output", "sovi_inputs.csv"))
counties.to_csv(here(path["ddpub"],'counties.csv'))
counties.to_file(here(path["ddpub"],'counties.gpkg'))
counties = pd.read_csv(here(path["ddpub"], "counties.csv"), dtype = {'GEOID': object})
# Select only the relevant columns
# Attribute name and expected influence on vulnerability
input_names = [['MEDAGE_ACS', 'pos', 'person', 'Median Age'],
['BLACK_ACS', 'pos', 'person', 'Pop African-American (%)'],
['QNATAM_ACS', 'pos', 'person', 'Pop Native American (%)'],
['QASIAN_ACS', 'pos', 'person', 'Pop Asian (%)'],
['QHISP_ACS', 'pos', 'person', 'Pop Hispanic (%)'],
['QAGEDEP_ACS', 'pos', 'person', 'Age Dependency (%)'],
['QPUNIT_ACS', 'pos', 'person', 'Persons Per Housing Unit'],
['PRENTER_ACS', 'pos', 'hu', 'Rental Housing (%)'],
['QNRRES_ACS', 'pos', 'person', 'Nursing Home Residents (%)'],
['QFEMALE_ACS', 'pos', 'person', 'Pop Female (%)'],
['QFHH_ACS', 'pos', 'hu', 'Female-Headed Households (%)'],
['QUNOCCHU_ACS', 'pos', 'hu', 'Vacant Housing (%)'],
['PERCAP_ALT', 'neg', 'person', 'Per-Capita Income'],
['QESL_ALT', 'pos', 'person', 'English as Second Language (%)'],
['QCVLUN', 'pos', 'person', 'Unemployment (%)'],
['QPOVTY', 'pos', 'person', 'Poverty (%)'],
['QMOHO', 'pos', 'hu', 'Mobile Homes (%)'],
['QED12LES_ALT', 'pos', 'person',
'Adults Completed <Grade 12 (%)'],
['QFEMLBR', 'pos', 'person', 'Female Employment (%)'],
['QEXTRCT_ALT', 'pos', 'person',
'Extractive Sector Employment (%)'],
['QSERV_ALT', 'pos', 'person', 'Service Sector Employment (%)'],
['QSSBEN', 'pos', 'hu', 'Social Security Income (%)'],
['QNOAUTO_ALT', 'pos', 'hu', 'No Automobile (%)'],
['QFAM', 'neg', 'person', 'Children in Married Families (%)'],
['QRICH200K', 'neg', 'hu', 'Annual Income >$200K (%)'],
['MDGRENT_ALT', 'neg', 'hu', 'Median Rent'],
['MHSEVAL_ALT', 'neg', 'hu', 'Median Home Value'],
['POPDENS', 'pos', 'person', 'Population Density']]
# Get attribute names
attr_names1 = [j[0] for j in input_names] + ['GEOID']
attr_names2 = [j[0] for j in input_names] + ['Geo_FIPS']
# Select only the columns needed to compute SoVI
counties = counties[attr_names1]
US_All = US_All[attr_names2]
counties["GEOID"] = "g" + counties["GEOID"]
counties['stateID'] = counties.GEOID.str.slice(0, 3, 1)
attr_names1.remove('GEOID')
counties = counties.set_index(counties["GEOID"]).sort_index()
US_All['stateID'] = US_All.Geo_FIPS.str.slice(0, 3, 1)
attr_names2.remove('Geo_FIPS')
US_All = US_All.set_index(US_All["Geo_FIPS"]).sort_index()
counties.eq(US_All).sum()
BLACK_ACS 3143 GEOID 0 Geo_FIPS 0 MDGRENT_ALT 3143 MEDAGE_ACS 3143 MHSEVAL_ALT 3143 PERCAP_ALT 3143 POPDENS 853 PRENTER_ACS 3143 QAGEDEP_ACS 3143 QASIAN_ACS 3143 QCVLUN 3143 QED12LES_ALT 3143 QESL_ALT 3143 QEXTRCT_ALT 3143 QFAM 3143 QFEMALE_ACS 3143 QFEMLBR 3143 QFHH_ACS 3143 QHISP_ACS 3143 QMOHO 3143 QNATAM_ACS 3143 QNOAUTO_ALT 3143 QNRRES_ACS 3143 QPOVTY 3143 QPUNIT_ACS 3143 QRICH200K 3143 QSERV_ALT 3143 QSSBEN 3143 QUNOCCHU_ACS 3143 stateID 3143 dtype: int64
Therre are 3143 observations in the dataset, so it appears that all variables match up perfectly except POPDENS. POPDENS is the one variable that was derived from the land area, so this was to be expected. Let us confirm that the differences between the two datasets are minor.
diff = (abs(counties[["POPDENS"]] - US_All[["POPDENS"]]))
print("Maximum difference:", diff.max()[0], "\nAverage difference:", diff.mean()[0])
Maximum difference: 0.949787960271351 Average difference: 0.0015009727346509663
With a maximum difference less than 1 and an average difference less than 0.01, our data is sufficiently close to Spielman et al.'s for our purposes.
Now we proceed to step P4, switching the directionality of variables as needed in order to ensure that higher values of a variable are associated with higher levels of vulnerability.
# Step P4
# Flip signs as needed to ensure that each variable contributes as expected to the final Sovi
for name, sign, sample, hrname in input_names:
if sign == 'neg':
counties[name] = -counties[name].values
print("Inverting variable:", name)
elif sign == 'pos':
pass
else:
print("problem in flipping signs")
raise
Inverting variable: PERCAP_ALT Inverting variable: QFAM Inverting variable: QRICH200K Inverting variable: MDGRENT_ALT Inverting variable: MHSEVAL_ALT
Spielman et al. constructed a class to conduct SPSS-style PCA with varimax rotation in python and validated their procedure against Cutter et al.'s SPSS workflow used to calculate SoVI. Below I include a workflow diagram that shows, without too much detail, the main operations and important outputs of their SPSS_PCA class. After that, I have included their relevant code.
class SPSS_PCA:
'''
A class that integrates most (all?) of the assumptions SPSS imbeds in their
implimnetation of principal components analysis (PCA), which can be found in
thier GUI under Analyze > Dimension Reduction > Factor. This class is not
intended to be a full blown recreation of the SPSS Factor Analysis GUI, but
it does replicate (possibly) the most common use cases. Note that this class
will not produce exactly the same results as SPSS, probably due to differences
in how eigenvectors/eigenvalues and/or singular values are computed. However,
this class does seem to get all the signs to match, which is not really necessary
but kinda nice. Most of the approach came from the official SPSS documentation.
References
----------
ftp://public.dhe.ibm.com/software/analytics/spss/documentation/statistics/20.0/en/client/Manuals/IBM_SPSS_Statistics_Algorithms.pdf
http://spssx-discussion.1045642.n5.nabble.com/Interpretation-of-PCA-td1074350.html
http://mdp-toolkit.sourceforge.net/api/mdp.nodes.WhiteningNode-class.html
https://github.com/mdp-toolkit/mdp-toolkit/blob/master/mdp/nodes/pca_nodes.py
Parameters
----------
inputs: numpy array
n x k numpy array; n observations and k variables on each observation
reduce: boolean (default=False)
If True, then use eigenvalues to determine which factors to keep; all
results will be based on just these factors. If False use all factors.
min_eig: float (default=1.0)
If reduce=True, then keep all factors with an eigenvalue greater than
min_eig. SPSS default is 1.0. If reduce=False, then min_eig is ignored.
varimax: boolean (default=False)
If True, then apply a varimax rotation to the results. If False, then
return the unrotated results only.
Attributes
----------
z_inputs: numpy array
z-scores of the input array.
comp_mat: numpy array
Component matrix (a.k.a, "loadings").
scores: numpy array
New uncorrelated vectors associated with each observation.
eigenvals_all: numpy array
Eigenvalues associated with each factor.
eigenvals: numpy array
Subset of eigenvalues_all reflecting only those that meet the
criterion defined by parameters reduce and min_eig.
weights: numpy array
Values applied to the input data (after z-scores) to get the PCA
scores. "Component score coefficient matrix" in SPSS or
"projection matrix" in the MDP library.
comms: numpy array
Communalities
sum_sq_load: numpy array
Sum of squared loadings.
comp_mat_rot: numpy array or None
Component matrix after rotation. Ordered from highest to lowest
variance explained based on sum_sq_load_rot. None if varimax=False.
scores_rot: numpy array or None
Uncorrelated vectors associated with each observation, after
rotation. None if varimax=False.
weights_rot: numpy array or None
Rotated values applied to the input data (after z-scores) to get
the PCA scores. None if varimax=False.
sum_sq_load_rot: numpy array or None
Sum of squared loadings for rotated results. None if
varimax=False.
'''
def __init__(self, inputs, reduce=False, min_eig=1.0, varimax=False):
# Step S1
z_inputs = ZSCORE(inputs) # seems necessary for SPSS "correlation matrix" setting (their default)
# The rest is step S2
# run base SPSS-style PCA to get all eigenvalues
pca_node = MDP.nodes.WhiteningNode() # settings for the PCA
scores = pca_node.execute(z_inputs) # base run PCA
eigenvalues_all = pca_node.d # rename PCA results
# run SPSS-style PCA based on user settings
pca_node = MDP.nodes.WhiteningNode(reduce=reduce, var_abs=min_eig) # settings for the PCA
scores = pca_node.execute(z_inputs) # run PCA (these have mean=0, std_dev=1)
weights = pca_node.v # rename PCA results (these might be a transformation of the eigenvectors)
eigenvalues = pca_node.d # rename PCA results
component_matrix = weights * eigenvalues # compute the loadings
component_matrix = self._reflect(component_matrix) # get signs to match SPSS
communalities = (component_matrix**2).sum(1) # compute the communalities
sum_sq_loadings = (component_matrix**2).sum(0) # note that this is the same as eigenvalues
weights_reflected = component_matrix/eigenvalues # get signs to match SPSS
scores_reflected = np.dot(z_inputs, weights_reflected) # note that abs(scores)=abs(scores_reflected)
if varimax:
# SPSS-style varimax rotation prep
c_normalizer = 1. / MDP.numx.sqrt(communalities) # used to normalize inputs to varimax
c_normalizer.shape = (component_matrix.shape[0],1) # reshape to vectorize normalization
cm_normalized = c_normalizer * component_matrix # normalize component matrix for varimax
# varimax rotation
cm_normalized_varimax = self._varimax(cm_normalized) # run varimax
c_normalizer2 = MDP.numx.sqrt(communalities) # used to denormalize varimax output
c_normalizer2.shape = (component_matrix.shape[0],1) # reshape to vectorize denormalization
cm_varimax = c_normalizer2 * cm_normalized_varimax # denormalize varimax output
# reorder varimax component matrix
sorter = (cm_varimax**2).sum(0) # base the ordering on sum of squared loadings
sorter = zip(sorter.tolist(), range(sorter.shape[0])) # add index to denote current order
sorter = sorted(sorter, key=itemgetter(0), reverse=True) # sort from largest to smallest
sum_sq_loadings_varimax, reorderer = zip(*sorter) # unzip the sorted list
sum_sq_loadings_varimax = np.array(sum_sq_loadings_varimax) # convert to array
cm_varimax = cm_varimax[:,reorderer] # reorder component matrix
# varimax scores
cm_varimax_reflected = self._reflect(cm_varimax) # get signs to match SPSS
# Step M4.5
variance_explained = pca.sum_sq_load_rot/sum(pca.eigenvals_all)
varimax_weights = np.dot(cm_varimax_reflected,
np.linalg.inv(np.dot(cm_varimax_reflected.T,
cm_varimax_reflected))) # CM(CM'CM)^-1
scores_varimax = np.dot(z_inputs, varimax_weights)
else:
comp_mat_rot = None
scores_rot = None
weights_rot = None
# assign output variables
self.z_inputs = z_inputs
self.scores = scores_reflected
self.comp_mat = component_matrix
self.eigenvals_all = eigenvalues_all
self.eigenvals = eigenvalues
self.weights = weights_reflected
self.comms = communalities
self.sum_sq_load = sum_sq_loadings
self.comp_mat_rot = cm_varimax_reflected
self.scores_rot = scores_varimax # PCA scores output
self.weights_rot = varimax_weights # PCA weights output
self.sum_sq_load_rot = sum_sq_loadings_varimax
def _reflect(self, cm):
# reflect factors with negative sums; SPSS default
cm = copy.deepcopy(cm)
reflector = cm.sum(0)
for column, measure in enumerate(reflector):
if measure < 0:
cm[:,column] = -cm[:,column]
return cm
def _varimax(self, Phi, gamma = 1.0, q = 100, tol = 1e-6):
# downloaded from http://en.wikipedia.org/wiki/Talk%3aVarimax_rotation
# also here http://stackoverflow.com/questions/17628589/perform-varimax-rotation-in-python-using-numpy
p,k = Phi.shape
R = np.eye(k)
d=0
for i in range(q):
d_old = d
Lambda = np.dot(Phi, R)
u,s,vh = np.linalg.svd(np.dot(Phi.T,np.asarray(Lambda)**3 - (gamma/p) *
np.dot(Lambda, np.diag(np.diag(np.dot(Lambda.T,Lambda))))))
R = np.dot(u,vh)
d = np.sum(s)
if d_old!=0 and d/d_old < 1 + tol:
break
return np.dot(Phi, R)
# Build FEMA subRegions Dict values= state ID's
FEMA_subs = dict()
FEMA_subs['FEMA_1'] = ['g23g33g25', 'g50', 'g09', 'g44']
FEMA_subs['FEMA_2'] = ['g36', 'g34']
FEMA_subs['FEMA_3'] = ['g42', 'g10', 'g11', 'g24', 'g51', 'g54']
FEMA_subs['FEMA_4'] = ['g21', 'g47', 'g37', 'g28', 'g01', 'g13', 'g45', 'g12']
FEMA_subs['FEMA_5'] = ['g27', 'g55', 'g26', 'g17', 'g18', 'g39']
FEMA_subs['FEMA_6'] = ['g35', 'g48', 'g40', 'g05', 'g22']
FEMA_subs['FEMA_7'] = ['g31', 'g19', 'g20', 'g29']
FEMA_subs['FEMA_8'] = ['g30', 'g38', 'g56', 'g46', 'g49', 'g08']
FEMA_subs['FEMA_9'] = ['g06', 'g32', 'g04']
FEMA_subs['FEMA_10'] = ['g53', 'g41', 'g16']
####################################
# DataFrames to hold US, FEMA region, and state level results
####################################
# Dict to hold variable loadings
# key will be [USA, Fema_region, stateid] depending on level of analysis
varContrib = {}
# National Score
US_Sovi_Score = pd.DataFrame(index=counties.GEOID,
columns=['sovi', 'rank'])
# In the FEMA_Region_Sovi_Score data frame ranks are BY FEMA REGION.
# The data frame holds both the SOVI score and the county rank
# This means that there should be 10 counties with rank 1 (one for each
# FEMA Region)
FEMA_Region_Sovi_Score = pd.DataFrame(index=counties.GEOID,
columns=['sovi', 'rank', 'fema_region'])
# Create New England conglomerate of states
# These are the FIPS codes for the states with the letter "g" appended
counties.loc[counties.stateID.isin(['g23', 'g33', 'g25']), 'stateID'] = 'g23g33g25'
# These are the states in the state level analysis
stateList = ['g23g33g25', 'g36', 'g51', 'g13',
'g17', 'g48', 'g29', 'g46', 'g06', 'g16']
# In the State_Sovi_Score data frame ranks are BY STATE.
# The data frame holds both the SOVI score and the county rank
# This means that there should be 10 counties with rank 1 (one for each
# state in stateList)
State_Sovi_Score = pd.DataFrame(
index=counties.index[counties.stateID.isin(stateList)],
columns=['sovi', 'rank', 'state_id'])
At this stage, we seek to calculate the SoVI ranks and variable weightings on the national, FEMA region, and state-level spatial extents. Below is a workflow for calculating SoVI, followed by the code for each spatial extent.
Step 4.5 is a new addition which calculates and prints a percent variance explained for each factor within the model.
#######################
# Compute National SoVI
#######################
# compute SoVI
# Step M2
inputData = counties.drop(['GEOID', 'stateID'], axis=1, inplace=False)
# Step M3
inputData_array = inputData.values # Convert DataFrame to NumPy array
# Step M4
pca = SPSS_PCA(inputData_array, reduce=True, varimax=True)
# Step M4.5
variance_explained = pca.sum_sq_load_rot/sum(pca.eigenvals_all)
print(variance_explained)
# Step M5
sovi_actual_us = pca.scores_rot.sum(1)
# Step M6
sovi_actual_us = pd.DataFrame(
sovi_actual_us, index=counties.GEOID, columns=['sovi'])
# print(sovi_actual_us)
# Step M8
sovi_actual_us['rank'] = sovi_actual_us.rank(
method='average', ascending=False)
US_Sovi_Score.update(sovi_actual_us)
# Step M9
attrib_contribution_us = pca.weights_rot.sum(1)
weights_rot_df = pd.DataFrame(pca.weights_rot, index = attr_names1)
print(weights_rot_df)
varContrib['USA'] = zip(attr_names1, attrib_contribution_us.tolist()) # Generate dictionary for all net loadings by variable for US
# Quick check of ranks: max should equal number of counties in US
try:
US_Sovi_Score['rank'].max() == len(counties)
except:
print("error in ranking check")
raise
# cleanup
del inputData
# del inputData_norm
del sovi_actual_us
del attrib_contribution_us
[0.17749433 0.1621382 0.12832775 0.09034435 0.06494221 0.05548509 0.05517805 0.05235274] 0 1 2 3 4 5 \ MEDAGE_ACS 0.028558 -0.097043 0.263585 -0.021732 0.064951 -0.047413 BLACK_ACS 0.231124 -0.081235 -0.032835 -0.128022 0.035158 -0.172086 QNATAM_ACS -0.120374 0.051713 0.034822 -0.061283 -0.125527 0.649159 QASIAN_ACS 0.059697 -0.177355 0.004209 0.044335 0.096726 -0.020888 QHISP_ACS -0.059316 0.021661 0.036007 0.430707 0.004060 -0.100344 QAGEDEP_ACS -0.032367 0.008297 0.231777 0.132433 0.137210 0.043627 QPUNIT_ACS -0.025747 0.041537 -0.025242 0.124587 -0.292591 0.168631 PRENTER_ACS 0.040047 0.009589 -0.232451 0.042722 0.268396 -0.015020 QNRRES_ACS -0.028948 0.071364 -0.074561 0.021579 0.395471 -0.046200 QFEMALE_ACS -0.029268 0.039311 0.079810 0.115028 -0.091273 0.036534 QFHH_ACS 0.144722 -0.003821 -0.043494 0.005639 -0.033313 0.016059 QUNOCCHU_ACS 0.007502 -0.099646 0.285825 0.023212 -0.113704 0.177186 PERCAP_ALT 0.033256 0.178163 -0.038022 0.052142 -0.033767 0.034974 QESL_ALT -0.018957 -0.042262 0.030951 0.386709 0.049780 -0.036214 QCVLUN 0.197523 -0.049893 0.087432 -0.044395 -0.101308 -0.093135 QPOVTY 0.123177 0.050919 -0.003049 0.039227 0.032417 0.060821 QMOHO 0.158646 -0.007006 0.098538 -0.064254 -0.184658 -0.079373 QED12LES_ALT 0.156469 0.022291 0.038671 0.155598 -0.000912 -0.104740 QFEMLBR -0.004034 0.001923 0.010762 -0.069927 -0.025838 0.005249 QEXTRCT_ALT -0.027330 0.053829 0.036552 0.118236 0.148751 0.117389 QSERV_ALT -0.049305 0.014880 0.040729 0.115068 0.068498 0.006772 QSSBEN 0.077827 -0.019041 0.244719 0.030591 0.020558 -0.038686 QNOAUTO_ALT 0.143955 -0.112268 0.029233 -0.053675 0.207615 0.352145 QFAM 0.147075 -0.021550 -0.012901 -0.026884 0.081235 -0.000337 QRICH200K -0.031148 0.213393 -0.076973 -0.012445 0.041138 0.006493 MDGRENT_ALT -0.012350 0.200516 -0.058312 -0.016243 0.143973 0.045744 MHSEVAL_ALT -0.019246 0.235548 -0.107857 0.000564 0.081574 -0.005632 POPDENS 0.162733 -0.206821 0.063250 0.000379 0.266078 0.132037 6 7 MEDAGE_ACS 0.041284 -0.020852 BLACK_ACS -0.031832 -0.086503 QNATAM_ACS 0.025288 0.040600 QASIAN_ACS -0.089701 -0.006796 QHISP_ACS 0.094955 0.134917 QAGEDEP_ACS 0.235682 0.003748 QPUNIT_ACS 0.109509 -0.084228 PRENTER_ACS -0.015465 0.069733 QNRRES_ACS -0.074874 0.067630 QFEMALE_ACS 0.704244 -0.117437 QFHH_ACS 0.107111 0.014858 QUNOCCHU_ACS -0.127205 0.126578 PERCAP_ALT 0.028228 0.024147 QESL_ALT 0.067625 0.043922 QCVLUN -0.127564 -0.018783 QPOVTY 0.015164 0.024725 QMOHO -0.104182 -0.184974 QED12LES_ALT 0.011764 -0.139073 QFEMLBR 0.257073 0.330415 QEXTRCT_ALT -0.091575 -0.241253 QSERV_ALT -0.155902 0.606872 QSSBEN 0.090472 -0.004187 QNOAUTO_ALT -0.013351 -0.154331 QFAM 0.010718 0.116535 QRICH200K -0.028958 0.103793 MDGRENT_ALT -0.001854 -0.075001 MHSEVAL_ALT 0.016087 -0.056065 POPDENS 0.005506 -0.268089
###########################
# Compute FEMA Region SoVI
###########################
for i in FEMA_subs:
# Step M1: Subset FEMA subregion
FEMARegionData = counties[counties['stateID'].isin(FEMA_subs[i])]
# Step M2
inputData = FEMARegionData.drop(
['GEOID', 'stateID'], axis=1, inplace=False)
# Step M3
inputData_array = inputData.values # Convert DataFrame to NumPy array
# Step M4
pca = SPSS_PCA(inputData_array, reduce=True, varimax=True)
# Step M5
sovi_actual_fema = pca.scores_rot.sum(1)
# Step M6
sovi_actual_fema = pd.DataFrame(
sovi_actual_fema, index=FEMARegionData.index, columns=['sovi'])
# Step M7
sovi_actual_fema['fema_region'] = i
# Step M8
sovi_actual_fema['rank'] = sovi_actual_fema['sovi'].rank(
method='average', ascending=False)
FEMA_Region_Sovi_Score.update(sovi_actual_fema)
# Step M9
attrib_contribution_fema = pca.weights_rot.sum(1)
varContrib[i] = zip(attr_names1, attrib_contribution_fema.tolist())
# cleanup
del FEMARegionData
del inputData
del sovi_actual_fema
del attrib_contribution_fema
############################
# Compute State Level SoVI
############################
for st in stateList:
# Step M1: Subset FEMA subregion
stateData = counties[counties.stateID == st]
# Step M2
inputData = stateData.drop(['GEOID', 'stateID'], axis=1, inplace=False)
# Step M3
inputData_array = inputData.values # Convert DataFrame to NumPy array
# Step M4
pca = SPSS_PCA(inputData_array, reduce=True, varimax=True)
# Step M5
sovi_actual = pca.scores_rot.sum(1)
# Step M6
sovi_actual = pd.DataFrame(
sovi_actual, index=stateData.index, columns=['sovi'])
# Step M7
sovi_actual['state_id'] = st
# Step M8
sovi_actual['rank'] = sovi_actual['sovi'].rank(
method='average', ascending=False)
State_Sovi_Score.update(sovi_actual)
# Step M9
attrib_contribution = pca.weights_rot.sum(1)
varContrib[st] = zip(attr_names1, attrib_contribution.tolist())
# cleanup
del stateData
del inputData
del sovi_actual
del attrib_contribution
Now that we have generated the SoVI scores for the 21 different models, we turn to our analysis of internal consistency.
This analysis checks for consistent SoVI rankings of counties in a region of interest (a state or group of small states) through three versions of a SoVI model, each using a different geographic extent for input data. Those extents are: 1) all counties in the country, 2) all the counties in a FEMA region, and 3) all counties in a single state or group of small states. The SoVI scores for the counties in the region of interest are selected and ranked. The agreement between the three sets of rankings is calculated using the Spearman's Rho rank correlation coefficient. If the model is internally consistent, one could expect a nearly perfect positive rank correlation close to 1, implying that counties have similar levels of social vulnerability vis a vis one another in the region of interest, regardless of how much extraneous information from other counties in the FEMA region or from the whole United States has been included in the SoVI model.
##########################################################################
# Ranks w/ Geographic Extent
# For each county rank within state for US, state, and fema_region sovis
##########################################################################
# Step IC1: Create an empty DataFrame with a column for each SoVI spatial extent and an index of each county FIPS code in the selected state(s) of the 10 FEMA regions
county_in_state_rank = pd.DataFrame(index=State_Sovi_Score.index,
columns=['state_sovi_rank', 'fema_region_sovi_rank', 'us_sovi_rank'])
for st in stateList:
if st == 'g23g33g25':
# Step IC2: Select the index and SoVI scores from national model for Maine, New Hampshire, and Massachusetts
st_cty_scores1 = US_Sovi_Score.loc[['g23' in s for s in US_Sovi_Score.index], 'sovi']
st_cty_scores2 = US_Sovi_Score.loc[['g33' in s for s in US_Sovi_Score.index], 'sovi']
st_cty_scores3 = US_Sovi_Score.loc[['g25' in s for s in US_Sovi_Score.index], 'sovi']
st_cty_scores = pd.concat([st_cty_scores1, st_cty_scores2, st_cty_scores3])
# Step IC3: Re-rank the national SoVI scores but just for the counties in the relevant states
county_in_state_rank.loc[st_cty_scores.index, 'us_sovi_rank'] = st_cty_scores.rank(method='average', ascending=False)
# Step IC4: Select the index and SoVI scores from FEMA model for Maine, New Hampshire, and Massachusetts
st_cty_scores1 = FEMA_Region_Sovi_Score.loc[['g23' in s for s in FEMA_Region_Sovi_Score.index], 'sovi']
st_cty_scores2 = FEMA_Region_Sovi_Score.loc[['g33' in s for s in FEMA_Region_Sovi_Score.index], 'sovi']
st_cty_scores3 = FEMA_Region_Sovi_Score.loc[['g25' in s for s in FEMA_Region_Sovi_Score.index], 'sovi']
st_cty_scores = pd.concat([st_cty_scores1, st_cty_scores2, st_cty_scores3])
# Step IC5: Re-rank the FEMA SoVI scores but just for the counties in the relevant states
county_in_state_rank.loc[st_cty_scores.index, 'fema_region_sovi_rank'] = st_cty_scores.rank(method='average', ascending=False)
# Step IC6: Pull the state-only SoVI ranks into the same dataframe as the other data
county_in_state_rank.loc[st_cty_scores.index, 'state_sovi_rank'] = State_Sovi_Score.loc[State_Sovi_Score['state_id'] == 'g23g33g25', 'rank']
else:
# Step IC2: select the index and SoVI scores from national model for the relevant state
st_cty_scores = US_Sovi_Score.loc[[st in s for s in US_Sovi_Score.index], 'sovi']
# Step IC3: Re-rank the national SoVI scores but just for the counties in the relevant state
county_in_state_rank.loc[st_cty_scores.index, 'us_sovi_rank'] = st_cty_scores.rank(method='average', ascending=False)
# Step IC4: Select the index and SoVI scores from FEMA model for the relevant state
st_cty_scores = FEMA_Region_Sovi_Score.loc[[st in s for s in FEMA_Region_Sovi_Score.index], 'sovi']
# Step IC5: Re-rank the FEMA SoVI scores but just for the counties in the relevant state
county_in_state_rank.loc[st_cty_scores.index, 'fema_region_sovi_rank'] = st_cty_scores.rank(method='average', ascending=False)
# Step IC6: Pull the state-only SoVI ranks into the same dataframe as the other data
county_in_state_rank.loc[st_cty_scores.index, 'state_sovi_rank'] = State_Sovi_Score.loc[State_Sovi_Score['state_id'] == st, 'rank']
######################
# CORRELATIONS
######################
# Step IC 7: Create an empty DataFrame to hold Spearman test results
state_corrs = pd.DataFrame(index = stateList, columns = ['spearman_r_st_fema', 'pvalue_st_fema', 'spearman_r_st_us', 'pvalue_st_us'])
for st in stateList:
if st == 'g23g33g25':
# Step IC8: Calculate spearman correlation between state and FEMA, state and national
multi_state_data_tmp1 = county_in_state_rank.loc[['g23' in s for s in county_in_state_rank.index], ]
multi_state_data_tmp2 = county_in_state_rank.loc[['g25' in s for s in county_in_state_rank.index], ]
multi_state_data_tmp3 = county_in_state_rank.loc[['g33' in s for s in county_in_state_rank.index], ]
multi_state_data_tmp = pd.concat([multi_state_data_tmp1, multi_state_data_tmp2, multi_state_data_tmp3])
st_fema_spearman = spearmanr(multi_state_data_tmp[['state_sovi_rank', 'fema_region_sovi_rank']])
st_us_spearman = spearmanr(multi_state_data_tmp[['state_sovi_rank', 'us_sovi_rank']])
state_corrs.loc['g23g33g25', ] = [st_fema_spearman[0], st_fema_spearman[1], st_us_spearman[0], st_us_spearman[1]]
else:
# Step IC8: Calculate spearman correlation between state and FEMA, state and national
st_fema_spearman = spearmanr(county_in_state_rank.loc[[st in s for s in county_in_state_rank.index], ['state_sovi_rank', 'fema_region_sovi_rank']])
st_us_spearman = spearmanr(county_in_state_rank.loc[[st in s for s in county_in_state_rank.index], ['state_sovi_rank', 'us_sovi_rank']])
state_corrs.loc[st, ] = [st_fema_spearman[0], st_fema_spearman[1], st_us_spearman[0], st_us_spearman[1]]
Finally, we investigate the questions surrounding theoretical consistency.
This analysis checks for consistent signs and ranks of variables across the same 21 models that were used in the internal consistency analysis. To evaluate the signs and ranks of variables, we sum all components together, producing one vector for each model containing the net effect of each variable on the SoVI score. Theoretical consistency is indicated by little variation amongst all models in the signs and magnitudes of variable weights. Theoretical inconsistency is indicated by substantial variation in the signs and weights of variables and by disagreement between a variable's theoretical influence and modeled influence on vulnerability.
Unplanned deviation: we were unable to find all of the code for this part the analysis, so we wrote much of this code ourselves.
# Step TC1: Create a DataFrame to hold variable contributions values
variable_contributions = pd.DataFrame(index=attr_names1)
# Step TC2: Add variable contributions values to DataFrame
for area in varContrib.keys():
variable_contributions[area] = [x for i, x in varContrib[area]]
# Step TC3: For all SoVI models, rank variables from the greatest to the least magnitudes
rankContrib = abs(variable_contributions).apply(rankdata, axis=0, method='average')
rankContrib = (28-rankContrib) + 1
# Step TC4: Sort variable rankings according to national model's most to least important
rankContrib = rankContrib.sort_values("USA", ascending = True).reset_index()
rankContrib.index = rankContrib["index"]
rankContrib = rankContrib.drop(columns = ["index"])
# Step TC5: Calculate summary statistics for each variable
summary_stats = pd.DataFrame( {"Min": rankContrib.min(axis = 1).round(),
"Max": rankContrib.max(axis = 1).round(),
"Range": rankContrib.max(axis = 1) - rankContrib.min(axis = 1).round(),
"Average": rankContrib.mean(axis = 1).round(2)
} )
# Step TC6: determine signs of USA model
def pos_neg(x):
if x > 0:
return "+"
else:
return "-"
usa = variable_contributions["USA"].apply(pos_neg)
# Step TC7: Determine all positive/negatives
reversals_adj = variable_contributions < 0
# Step TC8: Separate data
reference = reversals_adj[["USA"]]
other_vars = reversals_adj.drop(columns = ["USA"])
# Step TC9: Determine all reversals from expected sign
for i in range(len(other_vars.columns)):
other_vars.iloc[:, i] = reversals_adj["USA"].eq(other_vars.iloc[:, i]).eq(False)
# Step TC10: calculate reversals
reversal_sum = pd.DataFrame( {"Reversals": other_vars.sum(axis = 1)} )
# Step TC11: Join data
summary_stats = summary_stats.merge(reversal_sum, left_index = True, right_index = True)
# Step TC12: Join data
summary_stats = summary_stats.merge(usa, left_index = True, right_index = True)
# Step TC13: Change index labels to reflect any changes prior to SoVI calculation
for name, sign, sample, hrname in input_names:
if sign == 'neg':
summary_stats = summary_stats.rename(index={name: '- '+ name})
else:
pass
# Table aesthetics edits
summary_stats = summary_stats.rename(columns={"USA": "National Model"})
summary_stats = summary_stats.loc[:,['National Model', 'Reversals', 'Min', 'Average', 'Max', 'Range']]
We made a different choice than Spielman et al. in this table. Since we adjusted all variables such that larger values are theoretically associated with a higher degree of vulnerability before calculating SoVI, we would expect all outputs to be positive. In Spielman et al.'s "expected contribution" column, some signs were negative -- they recorded the directionality we would expect of the variables if they had made no adjustments before calculating SoVI. This leads to a misleading table because their "original contribution" column displays the signs of the model output including prior adjustments to directionaity, but the "expected contribution" column displays the signs if the model had not adjusted directionality. To make their "expected contribution" column consistent with their "original contribution" column, they would need all of the signs to be positive in the "expected contribution" column. We choose to simply not include an "expected contribution" column since there would be no variation within it anyway. Additionally, we add a negative sign in front of any variables that we changed the directionality of for the sake of clarity.
#####################################################
# OUTPUT TABLES
#####################################################
US_Sovi_Score.to_csv( here(path["ddpub"], 'US_Sovi_Score.csv') )
# In the FEMA_Region_Sovi_Score data frame ranks are BY FEMA REGION.
# The data frame holds both the SOVI score and the county rank
# This means that there should be 10 counties with rank 1 (one for each
# FEMA Region)
FEMA_Region_Sovi_Score.to_csv( here(path["ddpub"], 'FEMA_Region_Sovi_Score.csv') )
# In the State_Sovi_Score data frame ranks are BY STATE.
# The data frame holds both the SOVI score and the county rank
# This means that there should be 10 counties with rank 1 (one for each
# state in stateList)
State_Sovi_Score.to_csv( here(path["ddpub"], 'State_Sovi_Score.csv') )
# County rank within state for US, state, and fema_region sovis
county_in_state_rank.to_csv( here(path["ddpub"], 'County_in_State_Rank.csv') )
# Variable contributions for sovis at all geographic extents
variable_contributions.to_csv( here(path["ddpub"], 'variable_contributions.csv') )
# Correlation of ranks
state_corrs.to_csv( here(path["ddpub"], 'state_fema_us_rank_correlations.csv') )
First, we tested RPr-H1, that reproduced SoVI model scores for each county are not identical to the original study SoVI model scores for each county for each of the 21 SoVI models.
We define a function, check_it to check equivalency of the original output files to our reproduced output files.
def check_it(file, rounder = False):
'''
Given a file name, this function finds the corresponding file provided by Spielman et al. and the file produced
by our code and returns the number of matches for each column.
'''
global rpl
global og
global test
rpl = pd.read_csv( here(path["ddpub"], file) )
og = pd.read_csv( here(path["og_out"], file) )
og = og.rename(columns = {"Geo_FIPS": "GEOID"})
if "sovi" in rpl.columns:
rpl["sovi"] = rpl["sovi"].round(2)
og["sovi"] = og["sovi"].round(2)
if "Unnamed: 0" in rpl.columns:
rpl.index = rpl["Unnamed: 0"]
rpl = rpl.drop(columns = ["Unnamed: 0"])
if "Unnamed: 0" in og.columns:
og.index = og["Unnamed: 0"]
og = og.drop(columns = ["Unnamed: 0"])
if rounder != False:
og = og.round(rounder)
rpl = rpl.round(rounder)
test = rpl.eq(og)
if test.sum().eq(len(rpl)).sum() == len(test.sum()):
return print("All values match!")
else:
return test.sum()
check_it('US_Sovi_Score.csv')
GEOID 3143 sovi 3143 rank 3141 dtype: int64
merged = og.merge(rpl, how = "inner", on = "GEOID")
merged.loc[~test["rank"]]
GEOID | sovi_x | rank_x | sovi_y | rank_y | |
---|---|---|---|---|---|
1150 | g22075 | -4.53 | 2984.0 | -4.53 | 2983.0 |
3120 | g56001 | -4.53 | 2983.0 | -4.53 | 2984.0 |
We have identically reproduced national SoVI scores (rounded to 2 decimal points) for all 3143 counties compared to the original study, but two county ranks are different, probably due to the small differences between our area column and theirs.
check_it('FEMA_Region_Sovi_Score.csv')
GEOID 3143 sovi 3109 rank 3109 fema_region 3109 dtype: int64
Our check it function found potential differences in 34 counties. The following table shows the SoVI scores and ranks for those counties.
merged = og.merge(rpl, how = "inner", on = "GEOID")
merged.loc[~test["rank"] | ~test["sovi"] | ~test["fema_region"]]#.head()
GEOID | sovi_x | rank_x | fema_region_x | sovi_y | rank_y | fema_region_y | |
---|---|---|---|---|---|---|---|
67 | g02013 | NaN | NaN | NaN | NaN | NaN | NaN |
68 | g02016 | NaN | NaN | NaN | NaN | NaN | NaN |
69 | g02020 | NaN | NaN | NaN | NaN | NaN | NaN |
70 | g02050 | NaN | NaN | NaN | NaN | NaN | NaN |
71 | g02060 | NaN | NaN | NaN | NaN | NaN | NaN |
72 | g02068 | NaN | NaN | NaN | NaN | NaN | NaN |
73 | g02070 | NaN | NaN | NaN | NaN | NaN | NaN |
74 | g02090 | NaN | NaN | NaN | NaN | NaN | NaN |
75 | g02100 | NaN | NaN | NaN | NaN | NaN | NaN |
76 | g02105 | NaN | NaN | NaN | NaN | NaN | NaN |
77 | g02110 | NaN | NaN | NaN | NaN | NaN | NaN |
78 | g02122 | NaN | NaN | NaN | NaN | NaN | NaN |
79 | g02130 | NaN | NaN | NaN | NaN | NaN | NaN |
80 | g02150 | NaN | NaN | NaN | NaN | NaN | NaN |
81 | g02164 | NaN | NaN | NaN | NaN | NaN | NaN |
82 | g02170 | NaN | NaN | NaN | NaN | NaN | NaN |
83 | g02180 | NaN | NaN | NaN | NaN | NaN | NaN |
84 | g02185 | NaN | NaN | NaN | NaN | NaN | NaN |
85 | g02188 | NaN | NaN | NaN | NaN | NaN | NaN |
86 | g02195 | NaN | NaN | NaN | NaN | NaN | NaN |
87 | g02198 | NaN | NaN | NaN | NaN | NaN | NaN |
88 | g02220 | NaN | NaN | NaN | NaN | NaN | NaN |
89 | g02230 | NaN | NaN | NaN | NaN | NaN | NaN |
90 | g02240 | NaN | NaN | NaN | NaN | NaN | NaN |
91 | g02261 | NaN | NaN | NaN | NaN | NaN | NaN |
92 | g02270 | NaN | NaN | NaN | NaN | NaN | NaN |
93 | g02275 | NaN | NaN | NaN | NaN | NaN | NaN |
94 | g02282 | NaN | NaN | NaN | NaN | NaN | NaN |
95 | g02290 | NaN | NaN | NaN | NaN | NaN | NaN |
546 | g15001 | NaN | NaN | NaN | NaN | NaN | NaN |
547 | g15003 | NaN | NaN | NaN | NaN | NaN | NaN |
548 | g15005 | NaN | NaN | NaN | NaN | NaN | NaN |
549 | g15007 | NaN | NaN | NaN | NaN | NaN | NaN |
550 | g15009 | NaN | NaN | NaN | NaN | NaN | NaN |
These 34 counties are missing data in both the original study and our reproduction study. The counties and county equivalents are all located in Hawaii (FIPS code 15) and Alaska (FIPS code 02). In Spielman et al.'s code, when they define the states in FEMA region IX, they do not include HI, and when they define the states in FEMA region X, they do not include AK. All differences here arise from missing data in analogous places in both my output and theirs. This result was successfully reproduced.
check_it('State_Sovi_Score.csv')
All values match!
We have identically reproduced SoVI scores for all state models.
check_it("County_in_State_Rank.csv")
All values match!
We have identically reproduced the SoVI rankings in the state(s) of interest for all 21 models.
check_it("variable_contributions.csv", rounder = 3)
All values match!
When rounded to 3 decimal places, we have successfully reproduced all variable contributions for all models.
check_it("state_fema_us_rank_correlations.csv", rounder = 14)
All values match!
When rounded to 14 decimal places, we have succesfully reproduced all Spearman's rank correlations.
# Read files
counties = gpd.read_file( here(path["ddpub"], "counties.gpkg") )
USA = pd.read_csv( here(path["ddpub"], "US_Sovi_Score.csv") ).rename( columns={"sovi": "sovi_USA"} )
FEMA = pd.read_csv( here(path["ddpub"], "FEMA_Region_Sovi_Score.csv") ).rename( columns={"sovi": "sovi_FEMA"} )
CA = pd.read_csv( here(path["ddpub"], "State_Sovi_Score.csv") ).rename( columns={"sovi": "sovi_CA"} )
print(counties)
print(USA)
GEO_ID STATE COUNTY NAME LSAD CENSUSAREA GEOID \ 0 0500000US01029 01 029 Cleburne County 560.100 01029 1 0500000US01031 01 031 Coffee County 678.972 01031 2 0500000US01037 01 037 Coosa County 650.926 01037 3 0500000US01039 01 039 Covington County 1030.456 01039 4 0500000US01041 01 041 Crenshaw County 608.840 01041 ... ... ... ... ... ... ... ... 3138 0500000US56027 56 027 Niobrara County 2626.037 56027 3139 0500000US56031 56 031 Platte County 2084.208 56031 3140 0500000US56037 56 037 Sweetwater County 10426.649 56037 3141 0500000US56043 56 043 Washakie County 2238.549 56043 3142 0500000US56045 56 045 Weston County 2398.089 56045 B01002_001E B03002_001E B03002_004E ... QESL_ALT QED12LES_ALT \ 0 41.2 14905 573 ... 0.008884 0.257741 1 37.8 49966 8579 ... 0.034113 0.182346 2 44.4 11305 3558 ... 0.006690 0.242119 3 42.5 37803 4744 ... 0.009961 0.226478 4 40.3 13971 3099 ... 0.020430 0.236889 ... ... ... ... ... ... ... 3138 44.5 2478 21 ... 0.002096 0.099049 3139 47.8 8677 1 ... 0.014518 0.099858 3140 32.9 43890 191 ... 0.043683 0.097158 3141 41.6 8425 30 ... 0.031984 0.092084 3142 41.7 7152 4 ... 0.003431 0.088271 QEXTRCT_ALT QSERV_ALT QNOAUTO_ALT MDGRENT_ALT MHSEVAL_ALT \ 0 0.048569 0.158657 0.063791 566 91300.0 1 0.032847 0.155959 0.066835 638 126400.0 2 0.029628 0.165825 0.046578 573 74500.0 3 0.064493 0.155167 0.066179 560 86700.0 4 0.082556 0.140193 0.060907 488 70000.0 ... ... ... ... ... ... 3138 0.255245 0.178322 0.059230 557 122900.0 3139 0.156718 0.212128 0.039869 519 133300.0 3140 0.185796 0.151213 0.032235 881 174600.0 3141 0.127079 0.167287 0.045944 511 157700.0 3142 0.263604 0.157109 0.035835 640 131200.0 POPDENS MHSEVAL_ALT_LAG \ 0 26.611319 100212.500000 1 73.590664 86760.000000 2 17.367566 118200.000000 3 36.685700 107650.000000 4 22.946915 96983.333333 ... ... ... 3138 0.943627 132371.428571 3139 4.163212 163740.000000 3140 4.209406 225987.500000 3141 3.763599 175250.000000 3142 2.982375 165528.571429 geometry 0 POLYGON ((-85.38872 33.91304, -85.38088 33.873... 1 POLYGON ((-86.03044 31.61894, -86.00408 31.619... 2 POLYGON ((-86.00928 33.10164, -86.00917 33.090... 3 POLYGON ((-86.34851 30.99434, -86.35023 30.994... 4 POLYGON ((-86.14699 31.68045, -86.14711 31.663... ... ... 3138 POLYGON ((-104.05298 42.85955, -104.05286 42.7... 3139 POLYGON ((-104.77417 42.60996, -104.76422 42.6... 3140 POLYGON ((-109.05008 41.00066, -109.17368 41.0... 3141 POLYGON ((-107.12892 43.99455, -107.12797 43.9... 3142 POLYGON ((-104.05518 43.76163, -104.05514 43.7... [3143 rows x 82 columns] GEOID sovi_USA rank 0 g01001 -3.377903 2836.0 1 g01003 -1.181056 2189.0 2 g01005 -0.024258 1585.0 3 g01007 -4.418067 2970.0 4 g01009 -1.918604 2465.0 ... ... ... ... 3138 g56037 -6.339821 3100.0 3139 g56039 -5.187732 3035.0 3140 g56041 -3.962617 2929.0 3141 g56043 -0.352190 1789.0 3142 g56045 -3.634297 2874.0 [3143 rows x 3 columns]
# Edit counties GEOID to match other datasets
counties["GEOID"] = "g" + counties["GEOID"]
# Select just the rows and columns needed
counties_CA = counties.loc[counties["STATE"] == "06"]
counties_CA = counties_CA[["GEOID", "geometry"]]
# Join all datasets
counties_CA = counties_CA.merge(USA, on = "GEOID")
counties_CA = counties_CA.merge(FEMA, on = "GEOID")
counties_CA = counties_CA.merge(CA, on = "GEOID")
counties_CA['rank_USA'] = counties_CA['sovi_USA'].rank(method='average', ascending=False)
counties_CA['rank_FEMA'] = counties_CA['sovi_FEMA'].rank(method='average', ascending=False)
counties_CA['rank_CA'] = counties_CA['sovi_CA'].rank(method='average', ascending=False)
mycolor = ListedColormap('#DBDBDB')
# Create map of national SoVI scores by county
sovi_usa = counties.merge(USA, on = "GEOID")
print(sovi_usa)
# Select just the rows and columns needed
sovi_usa.loc[counties["STATE"] != "02"].loc[counties["STATE"] != "15"]
#counties_CA = counties_CA[["GEOID", "geometry"]]
GEO_ID STATE COUNTY NAME LSAD CENSUSAREA GEOID \ 0 0500000US01029 01 029 Cleburne County 560.100 g01029 1 0500000US01031 01 031 Coffee County 678.972 g01031 2 0500000US01037 01 037 Coosa County 650.926 g01037 3 0500000US01039 01 039 Covington County 1030.456 g01039 4 0500000US01041 01 041 Crenshaw County 608.840 g01041 ... ... ... ... ... ... ... ... 3138 0500000US56027 56 027 Niobrara County 2626.037 g56027 3139 0500000US56031 56 031 Platte County 2084.208 g56031 3140 0500000US56037 56 037 Sweetwater County 10426.649 g56037 3141 0500000US56043 56 043 Washakie County 2238.549 g56043 3142 0500000US56045 56 045 Weston County 2398.089 g56045 B01002_001E B03002_001E B03002_004E ... QEXTRCT_ALT QSERV_ALT \ 0 41.2 14905 573 ... 0.048569 0.158657 1 37.8 49966 8579 ... 0.032847 0.155959 2 44.4 11305 3558 ... 0.029628 0.165825 3 42.5 37803 4744 ... 0.064493 0.155167 4 40.3 13971 3099 ... 0.082556 0.140193 ... ... ... ... ... ... ... 3138 44.5 2478 21 ... 0.255245 0.178322 3139 47.8 8677 1 ... 0.156718 0.212128 3140 32.9 43890 191 ... 0.185796 0.151213 3141 41.6 8425 30 ... 0.127079 0.167287 3142 41.7 7152 4 ... 0.263604 0.157109 QNOAUTO_ALT MDGRENT_ALT MHSEVAL_ALT POPDENS MHSEVAL_ALT_LAG \ 0 0.063791 566 91300.0 26.611319 100212.500000 1 0.066835 638 126400.0 73.590664 86760.000000 2 0.046578 573 74500.0 17.367566 118200.000000 3 0.066179 560 86700.0 36.685700 107650.000000 4 0.060907 488 70000.0 22.946915 96983.333333 ... ... ... ... ... ... 3138 0.059230 557 122900.0 0.943627 132371.428571 3139 0.039869 519 133300.0 4.163212 163740.000000 3140 0.032235 881 174600.0 4.209406 225987.500000 3141 0.045944 511 157700.0 3.763599 175250.000000 3142 0.035835 640 131200.0 2.982375 165528.571429 geometry sovi_USA rank 0 POLYGON ((-85.38872 33.91304, -85.38088 33.873... -0.620170 1929.0 1 POLYGON ((-86.03044 31.61894, -86.00408 31.619... -0.812562 2020.0 2 POLYGON ((-86.00928 33.10164, -86.00917 33.090... -1.226775 2218.0 3 POLYGON ((-86.34851 30.99434, -86.35023 30.994... 1.536396 796.0 4 POLYGON ((-86.14699 31.68045, -86.14711 31.663... -0.451828 1850.0 ... ... ... ... 3138 POLYGON ((-104.05298 42.85955, -104.05286 42.7... 1.234826 916.0 3139 POLYGON ((-104.77417 42.60996, -104.76422 42.6... 1.325057 885.0 3140 POLYGON ((-109.05008 41.00066, -109.17368 41.0... -6.339821 3100.0 3141 POLYGON ((-107.12892 43.99455, -107.12797 43.9... -0.352190 1789.0 3142 POLYGON ((-104.05518 43.76163, -104.05514 43.7... -3.634297 2874.0 [3143 rows x 84 columns]
GEO_ID | STATE | COUNTY | NAME | LSAD | CENSUSAREA | GEOID | B01002_001E | B03002_001E | B03002_004E | ... | QEXTRCT_ALT | QSERV_ALT | QNOAUTO_ALT | MDGRENT_ALT | MHSEVAL_ALT | POPDENS | MHSEVAL_ALT_LAG | geometry | sovi_USA | rank | |
---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|---|
0 | 0500000US01029 | 01 | 029 | Cleburne | County | 560.100 | g01029 | 41.2 | 14905 | 573 | ... | 0.048569 | 0.158657 | 0.063791 | 566 | 91300.0 | 26.611319 | 100212.500000 | POLYGON ((-85.38872 33.91304, -85.38088 33.873... | -0.620170 | 1929.0 |
1 | 0500000US01031 | 01 | 031 | Coffee | County | 678.972 | g01031 | 37.8 | 49966 | 8579 | ... | 0.032847 | 0.155959 | 0.066835 | 638 | 126400.0 | 73.590664 | 86760.000000 | POLYGON ((-86.03044 31.61894, -86.00408 31.619... | -0.812562 | 2020.0 |
2 | 0500000US01037 | 01 | 037 | Coosa | County | 650.926 | g01037 | 44.4 | 11305 | 3558 | ... | 0.029628 | 0.165825 | 0.046578 | 573 | 74500.0 | 17.367566 | 118200.000000 | POLYGON ((-86.00928 33.10164, -86.00917 33.090... | -1.226775 | 2218.0 |
3 | 0500000US01039 | 01 | 039 | Covington | County | 1030.456 | g01039 | 42.5 | 37803 | 4744 | ... | 0.064493 | 0.155167 | 0.066179 | 560 | 86700.0 | 36.685700 | 107650.000000 | POLYGON ((-86.34851 30.99434, -86.35023 30.994... | 1.536396 | 796.0 |
4 | 0500000US01041 | 01 | 041 | Crenshaw | County | 608.840 | g01041 | 40.3 | 13971 | 3099 | ... | 0.082556 | 0.140193 | 0.060907 | 488 | 70000.0 | 22.946915 | 96983.333333 | POLYGON ((-86.14699 31.68045, -86.14711 31.663... | -0.451828 | 1850.0 |
... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... | ... |
3138 | 0500000US56027 | 56 | 027 | Niobrara | County | 2626.037 | g56027 | 44.5 | 2478 | 21 | ... | 0.255245 | 0.178322 | 0.059230 | 557 | 122900.0 | 0.943627 | 132371.428571 | POLYGON ((-104.05298 42.85955, -104.05286 42.7... | 1.234826 | 916.0 |
3139 | 0500000US56031 | 56 | 031 | Platte | County | 2084.208 | g56031 | 47.8 | 8677 | 1 | ... | 0.156718 | 0.212128 | 0.039869 | 519 | 133300.0 | 4.163212 | 163740.000000 | POLYGON ((-104.77417 42.60996, -104.76422 42.6... | 1.325057 | 885.0 |
3140 | 0500000US56037 | 56 | 037 | Sweetwater | County | 10426.649 | g56037 | 32.9 | 43890 | 191 | ... | 0.185796 | 0.151213 | 0.032235 | 881 | 174600.0 | 4.209406 | 225987.500000 | POLYGON ((-109.05008 41.00066, -109.17368 41.0... | -6.339821 | 3100.0 |
3141 | 0500000US56043 | 56 | 043 | Washakie | County | 2238.549 | g56043 | 41.6 | 8425 | 30 | ... | 0.127079 | 0.167287 | 0.045944 | 511 | 157700.0 | 3.763599 | 175250.000000 | POLYGON ((-107.12892 43.99455, -107.12797 43.9... | -0.352190 | 1789.0 |
3142 | 0500000US56045 | 56 | 045 | Weston | County | 2398.089 | g56045 | 41.7 | 7152 | 4 | ... | 0.263604 | 0.157109 | 0.035835 | 640 | 131200.0 | 2.982375 | 165528.571429 | POLYGON ((-104.05518 43.76163, -104.05514 43.7... | -3.634297 | 2874.0 |
3109 rows × 84 columns
# fig, ax = plt.subplots(figsize=(50, 50))
# fig.tight_layout()
# ax.set_title("SoVI Score by County")
# sovi_usa.plot(ax = ax, cmap = "Blues", edgecolor = 'black', linewidth = .1, column = "sovi_usa")
# sovi_usa
Next, we visualized the national SoVI scores. We had not previously mapped the social vulnerability index nationwide. The following interactive map shows the SoVI score of every county in the US.
map = folium.Map(location=[39, -98], tiles="CartoDB Positron", zoom_start=4)
folium.Choropleth(
geo_data=counties,
name="choropleth",
data=USA,
columns=["GEOID","sovi_USA"],
key_on="feature.properties.GEOID",
fill_color="YlGn",
fill_opacity=0.7,
line_opacity=0.2,
legend_name="SoVI",
).add_to(map)
folium.LayerControl().add_to(map)
map