
#Install apache spark on ec2 ubuntu download#
Example data Download and run example notebook blocks You should see the Jupyter admin page.Ĭreate two directories 'data_in' and 'data_out' by clicking New > Folder.Ĭreate a new notebook using Python 3 or download the example notebook.Īnd download the example data to the 'data_in' directory. The problem is that sometimes our data is not in proper format, our Glue crawler or job fails, and we stare blankly at an insufficient error log in CloudWatch. It is very useful for transforming data from CSV to Parquet for later use in AWS Athena, or loading large flat files into an RDBMS as part of other processes.

In the research programming group we have begun using AWS Glue for several projects. We have not tested it, but it may be preferable if you like using containers.ĭeveloping AWS Glue ETL jobs locally using a container Introduction Since our original publishing of this How-To, AWS has created their own documentation using Docker containers. It may not be worth the trouble, but if you have had as many issues with Glue that we have, it could save you time in the long run. Note: This involves many steps and correcting several things manually.
