]> code.communitydata.science - cdsc_reddit.git/commitdiff
Create a second dataset sorted by author.
authorNate E TeBlunthuis <nathante@n2347.hyak.local>
Mon, 6 Jul 2020 06:24:40 +0000 (23:24 -0700)
committerNate E TeBlunthuis <nathante@n2347.hyak.local>
Mon, 6 Jul 2020 06:27:05 +0000 (23:27 -0700)
comments_2_parquet.py

index 3042f58efae5e4d0545fb7731fccdea0499c23ed..cff16342dede2f0d04f71aeb36d0c4e1f21ab999 100755 (executable)
@@ -1,7 +1,7 @@
 
 #!/usr/bin/env python3
 import pyspark
-from pyspark.sql import functions as f
+nfrom pyspark.sql import functions as f
 from pyspark.sql.types import *
 from pyspark import SparkConf, SparkContext
 from pyspark.sql import SparkSession, SQLContext
@@ -128,5 +128,8 @@ df = df.withColumn("Month",f.month(f.col("CreatedAt")))
 df = df.withColumn("Year",f.year(f.col("CreatedAt")))
 df = df.withColumn("Day",f.dayofmonth(f.col("CreatedAt")))
 df = df.withColumn("subreddit_hash",f.sha2(f.col("subreddit"), 256)[0:3])
-df = df.sort(["subreddit","author","link_id","parent_id","Year","Month","Day"],ascending=True)
-df.write.parquet("/gscratch/comdata/output/reddit_comments.parquet", partitionBy=["Year",'Month'],mode='overwrite')
+df2 = df.sort(["subreddit","author","link_id","parent_id","Year","Month","Day"],ascending=True)
+df2.write.parquet("/gscratch/comdata/output/reddit_comments_by_subreddit.parquet", partitionBy=["Year",'Month'],mode='overwrite')
+
+df3 = df.sort(["author","CreatetdAt","subreddit","link_id","parent_id","Year","Month","Day"],ascending=True)
+df3.write.parquet("/gscratch/comdata/output/reddit_comments_by_author.parquet", partitionBy=["Year",'Month'],mode='overwrite')

Community Data Science Collective || Want to submit a patch?