Skip to content

Commit 8445c84

Browse files
authored
Merge pull request #2 from mansueto-institute/updates-prepub1
Updates prepub1
2 parents c1860eb + 1d839fc commit 8445c84

10 files changed

+3060
-77
lines changed

.gitignore

Lines changed: 6 additions & 0 deletions
Original file line numberDiff line numberDiff line change
@@ -2,5 +2,11 @@
22
.vscode
33
.DS_Store
44
__pycache__/
5+
/__pycache__
56
.*
67
!.gitignore
8+
/data/complexity-analysis
9+
/data/dhs-analysis
10+
/data/graph-viz
11+
/data/africa_data.parquet
12+
/data/africa_geodata.parquet

complexity-analysis.R

Lines changed: 19 additions & 15 deletions
Original file line numberDiff line numberDiff line change
@@ -20,7 +20,7 @@ library(viridis)
2020
library(scatterpie)
2121
library(ggrepel)
2222
library(DescTools)
23-
options(scipen = 999)
23+
options(scipen = 9999)
2424
options(lifecycle_verbosity = "warning")
2525

2626
# Load aggregation function
@@ -29,18 +29,18 @@ getwd()
2929
source('aggregation_func.R') # should be in the ~/kblock-analysis directory
3030

3131
# Read in data ------------------------------------------------------------
32-
dir.create('complexity-analysis')
33-
dir.create('complexity-analysis/input-data')
34-
wd_input = 'complexity-analysis/input-data'
3532
# Download block level data from this URL: dsbprylw7ncuq.cloudfront.net/AF/africa_data.parquet
36-
curl::multi_download("dsbprylw7ncuq.cloudfront.net/AF/africa_data.parquet", "complexity-analysis/input-data/africa_data.parquet", resume = TRUE)
37-
df_combined <- read_parquet(paste0(wd_input,'/africa_data.parquet'))
38-
33+
if (!file.exists(paste0("data/africa_data.parquet"))) {
34+
curl::multi_download("dsbprylw7ncuq.cloudfront.net/AF/africa_data.parquet", "complexity-analysis/input-data/africa_data.parquet", resume = TRUE)
35+
} else {
36+
df_combined <- read_parquet(paste0('data/africa_data.parquet'))
37+
}
38+
3939
# Create output directories
40-
dir.create('complexity-analysis/output-data')
41-
dir.create('complexity-analysis/output-data/data')
42-
dir.create('complexity-analysis/output-data/viz')
43-
wd_output = 'complexity-analysis/output-data'
40+
dir.create('data/complexity-analysis')
41+
dir.create('data/complexity-analysis/data')
42+
dir.create('data/complexity-analysis/viz')
43+
wd_output = 'data/complexity-analysis'
4444

4545
# K-complexity groupings --------------------------------------------------
4646

@@ -954,7 +954,7 @@ rm(a, b, conurban_bars, conurban_bars_data)
954954

955955
# Inequality charts -------------------------------------------------------
956956

957-
urban_ineq <- read_parquet(paste0(wd_input,'/africa_data.parquet')) %>%
957+
urban_ineq <- read_parquet(paste0('data/africa_data.parquet')) %>%
958958
filter(area_type == 'Urban')
959959

960960
names(urban_ineq)
@@ -1039,9 +1039,13 @@ rm(conurban_sd, urban_sd, agglos_sd, scatter_gini, scatter_sd)
10391039

10401040
# -------------------------------------------------------------------------
10411041
# Download from this URL: dsbprylw7ncuq.cloudfront.net/AF/africa_geodata.parquet
1042-
curl::multi_download("dsbprylw7ncuq.cloudfront.net/AF/africa_geodata.parquet", "complexity-analysis/input-data/africa_geodata.parquet", resume = TRUE)
1042+
if (!file.exists(paste0("data/africa_geodata.parquet"))) {
1043+
curl::multi_download("dsbprylw7ncuq.cloudfront.net/AF/africa_geodata.parquet", "data/africa_geodata.parquet", resume = TRUE)
1044+
} else {
1045+
print('africa_geodata.parquet already downloaded.')
1046+
}
10431047

1044-
area_data <- arrow::open_dataset(paste0(wd_input,'/africa_geodata.parquet')) %>%
1048+
area_data <- arrow::open_dataset(paste0('data/africa_geodata.parquet')) %>%
10451049
filter(urban_id %in% c('ghsl_3798','periurban_925')) %>%
10461050
read_sf_dataset() %>%
10471051
st_set_crs(4326) %>%
@@ -1257,7 +1261,7 @@ ggsave(plot = layers_viz, filename = paste0(wd_output,'/viz/zoom_maps.pdf'),
12571261

12581262
# Maps --------------------------------------------------------------------
12591263

1260-
area_data <- arrow::open_dataset(paste0(wd_input,'/africa_geodata.parquet')) %>%
1264+
area_data <- arrow::open_dataset(paste0('data/africa_geodata.parquet')) %>%
12611265
filter(urban_id %in% c('ghsl_2125')) %>%
12621266
read_sf_dataset() %>%
12631267
st_set_crs(4326) %>%

data/ghsl_to_dhs.parquet

95.1 KB
Binary file not shown.

data/streets_dhs_regions.csv

Lines changed: 376 additions & 0 deletions
Large diffs are not rendered by default.

data/streets_metrics.parquet

559 KB
Binary file not shown.

0 commit comments

Comments
 (0)