PySpark and boto3 driver to migrate telemetry data over to DynamoDB
Перейти к файлу
Victor Ng e41df7e1f1 pile-o logging and increased the sample size by 100x 2018-02-06 09:41:03 -05:00
deps added wheel dependencies for argparse and boto3 so that we can safely 2018-01-30 14:38:19 -05:00
scripts removed the call to push boto3 to spark workers 2018-02-01 21:16:46 -05:00
taar_loader pile-o logging and increased the sample size by 100x 2018-02-06 09:41:03 -05:00
.gitignore init commit. 2018-01-30 12:38:39 -05:00
LICENSE init commit. 2018-01-30 12:38:39 -05:00
Makefile added extra JSON serialization safeties 2018-02-06 09:27:51 -05:00
README.md added readme 2018-02-01 20:57:27 -05:00
setup.cfg added more metadata to support PyPI 2018-02-01 21:14:13 -05:00
setup.py added more metadata to support PyPI 2018-02-01 21:14:13 -05:00

README.md

taar_loader

This library provides basic support to run ETL jobs to migrate data from Parquet files stored in S3 and copy the data into a DynamoDB instance for the TAAR project.