-#!/usr/bin/env bash
+## this should be run manually since we don't have a nice way to wait on parallel_sql jobs
-# part2 should be run on one ore more spark nodes
+#!/usr/bin/env bash
-./submissions_2_parquet_part1.py
+./parse_submissions.sh
start_spark_and_run.sh 1 $(pwd)/submissions_2_parquet_part2.py