Skip to content

Instantly share code, notes, and snippets.

@rclark
Last active February 1, 2016 23:38
Show Gist options
  • Star 0 You must be signed in to star a gist
  • Fork 0 You must be signed in to fork a gist
  • Save rclark/a396954c3a27a17134e1 to your computer and use it in GitHub Desktop.
Save rclark/a396954c3a27a17134e1 to your computer and use it in GitHub Desktop.
      |                                           |                                       |                                                     |                                                         |
      |                                           |                                       |                                                     |                                                         |
      |--------- really no time at all -----------|------------ 7 to 14 days -------------|-------------- really no time at all ----------------|----------------- shouldn't be too long -----------------|
      |                                           |                                       |                                                     |                                                         |
      |                                           |                                       |                                                     |                                                         |
Start stack in backfill mode:                 Start backfilling dynamodb:              Backfill is done:                                      Feed changed feature ids down Dynamosm pipeline:          All done!
- changes begin accumulating in dynamodb      - chop up a pbf, ingest it              - switch out of backfill mode                           - Dynamosm pipeline will only ever make GeoJSON with      - keep on chugging through ongoing changes
- no changes in geojson cache                 - convert the pbf to geojson and        - new changes start landing new geojson on S3             the current version of data in dynamodb, so this        - reduce table capacities to normal levels
- increase table capacity                       write it all to S3                    - gather change XML since pbf creation into a file        won't run the risk of overwriting newer changes
                                                                                      - parse and dedupe to get ids of features that changed  - There will not be an inordinate number of things
                                                                                                                                              - Table throughputs can still be high from backfill
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment