mirror of
https://github.com/a-b-street/abstreet.git
synced 2024-12-24 23:15:24 +03:00
Polish up the cloud importer, based on its first real use. #326
- Give the worker assigned Seattle a much bigger machine type - Create a script to grab the results from S3 and finalize them
This commit is contained in:
parent
e876998958
commit
f5dcd9bfff
21
cloud/gather_output.sh
Executable file
21
cloud/gather_output.sh
Executable file
@ -0,0 +1,21 @@
|
||||
#!/bin/bash
|
||||
# This script grabs all files changed by VMs, copying them locally.
|
||||
#
|
||||
# This process is only runnable by Dustin, due to current S3 permissions.
|
||||
#
|
||||
# Run from the repo's root dir: cloud/gather_output.sh
|
||||
|
||||
set -e
|
||||
set -x
|
||||
|
||||
EXPERIMENT_TAG=$1
|
||||
if [ "$EXPERIMENT_TAG" == "" ]; then
|
||||
echo Missing args;
|
||||
exit 1;
|
||||
fi
|
||||
|
||||
aws s3 cp --recursive s3://abstreet/$EXPERIMENT_TAG/data/ data/
|
||||
# gunzip all of the changed files, overwriting the local copies
|
||||
find data/ -path data/system/assets -prune -o -name '*.gz' -print -exec gunzip -f '{}' ';'
|
||||
|
||||
echo "Done! Validate the files, run updater --upload as usual, and don't forget to clean up s3://abstreet/$EXPERIMENT_TAG"
|
@ -3,7 +3,7 @@
|
||||
# creates a bunch of GCE VMs, and runs the importer there on all cities, using
|
||||
# static sharding.
|
||||
#
|
||||
# This process is only runnable by Dustin, due to current GCE/EC2 permissions.
|
||||
# This process is only runnable by Dustin, due to current GCE/S3 permissions.
|
||||
#
|
||||
# Run from the repo's root dir: cloud/start_batch_import.sh
|
||||
|
||||
@ -20,10 +20,12 @@ NUM_WORKERS=10
|
||||
ZONE=us-east1-b
|
||||
# See other options: https://cloud.google.com/compute/docs/machine-types
|
||||
# Particularly... e2-standard-2, n2-standard-2, c2-standard-4
|
||||
MACHINE_TYPE=e2-standard-2
|
||||
SMALL_MACHINE_TYPE=e2-standard-2
|
||||
LARGE_MACHINE_TYPE=c2-standard-4
|
||||
# All of data/ is currently around 30GB
|
||||
DISK_SIZE=40GB
|
||||
# Compressing and checksumming gigantic files needs more IOPS
|
||||
# TODO But wait, e2-standard-2 doesn't support local PD?!
|
||||
DISK_TYPE=pd-ssd
|
||||
# Haha, using a project from college, my last traffic sim...
|
||||
PROJECT=aorta-routes
|
||||
@ -64,11 +66,19 @@ function create_vms {
|
||||
# gcloud integration tests...
|
||||
# https://issuetracker.google.com/issues/188462253
|
||||
for ((i = 0; i < $NUM_WORKERS; i++)); do
|
||||
# The first shard always handles Seattle, which needs more than
|
||||
# 8GB of memory. Just give it really hefty hardware.
|
||||
if [ $i == 0 ]; then
|
||||
machine_type=$LARGE_MACHINE_TYPE;
|
||||
else
|
||||
machine_type=$SMALL_MACHINE_TYPE;
|
||||
fi
|
||||
|
||||
gcloud compute \
|
||||
--project=$PROJECT \
|
||||
instances create "worker-$i" \
|
||||
--zone=$ZONE \
|
||||
--machine-type=$MACHINE_TYPE \
|
||||
--machine-type=$machine_type \
|
||||
--boot-disk-size=$DISK_SIZE \
|
||||
--boot-disk-type=$DISK_TYPE \
|
||||
--image-family=ubuntu-2004-lts \
|
||||
|
Loading…
Reference in New Issue
Block a user