From 67857a3b05b9bf7db684b5d2ab34802fa02ab08f Mon Sep 17 00:00:00 2001 From: Nate E TeBlunthuis Date: Sun, 5 Jul 2020 23:24:40 -0700 Subject: [PATCH] Create a second dataset sorted by author. --- comments_2_parquet.py | 9 ++++++--- 1 file changed, 6 insertions(+), 3 deletions(-) diff --git a/comments_2_parquet.py b/comments_2_parquet.py index 3042f58..cff1634 100755 --- a/comments_2_parquet.py +++ b/comments_2_parquet.py @@ -1,7 +1,7 @@ #!/usr/bin/env python3 import pyspark -from pyspark.sql import functions as f +nfrom pyspark.sql import functions as f from pyspark.sql.types import * from pyspark import SparkConf, SparkContext from pyspark.sql import SparkSession, SQLContext @@ -128,5 +128,8 @@ df = df.withColumn("Month",f.month(f.col("CreatedAt"))) df = df.withColumn("Year",f.year(f.col("CreatedAt"))) df = df.withColumn("Day",f.dayofmonth(f.col("CreatedAt"))) df = df.withColumn("subreddit_hash",f.sha2(f.col("subreddit"), 256)[0:3]) -df = df.sort(["subreddit","author","link_id","parent_id","Year","Month","Day"],ascending=True) -df.write.parquet("/gscratch/comdata/output/reddit_comments.parquet", partitionBy=["Year",'Month'],mode='overwrite') +df2 = df.sort(["subreddit","author","link_id","parent_id","Year","Month","Day"],ascending=True) +df2.write.parquet("/gscratch/comdata/output/reddit_comments_by_subreddit.parquet", partitionBy=["Year",'Month'],mode='overwrite') + +df3 = df.sort(["author","CreatetdAt","subreddit","link_id","parent_id","Year","Month","Day"],ascending=True) +df3.write.parquet("/gscratch/comdata/output/reddit_comments_by_author.parquet", partitionBy=["Year",'Month'],mode='overwrite') -- 2.39.2