diff --git a/eda/feededa/src/main/scala/com/gmei/GmeiConfig.scala b/eda/feededa/src/main/scala/com/gmei/GmeiConfig.scala index 0c62473d33685fe44872a8a30a1a3547f850fe16..830795d3f7ab892632a68ab435cac71522a35c8c 100644 --- a/eda/feededa/src/main/scala/com/gmei/GmeiConfig.scala +++ b/eda/feededa/src/main/scala/com/gmei/GmeiConfig.scala @@ -48,6 +48,7 @@ object GmeiConfig extends Serializable { .builder() .config(sparkConf) .appName("feededa") + .enableHiveSupport() .getOrCreate() val context = SparkContext.getOrCreate(sparkConf) diff --git a/eda/feededa/src/main/scala/com/gmei/WeafareStat.scala b/eda/feededa/src/main/scala/com/gmei/WeafareStat.scala index bb6b8bedcb5fc7849208ee043d7e2fc88f6aa242..87e0507cc62945a2b9f0d012612730cb98a5a6f4 100644 --- a/eda/feededa/src/main/scala/com/gmei/WeafareStat.scala +++ b/eda/feededa/src/main/scala/com/gmei/WeafareStat.scala @@ -81,7 +81,9 @@ object WeafareStat { ) video_count.show() val output1 = "./test_vd_cids.csv" - video_count.repartition(1).write.format("com.databricks.spark.csv").option("header","true").save(output1) + video_count.repartition(1) + .write.format("com.databricks.spark.csv") + .option("header","true").save(output1) val txt_count = sc.sql( s""" @@ -95,7 +97,9 @@ object WeafareStat { ) txt_count.show() val output2 = "./test_txt_cids.csv" - txt_count.repartition(1).write.format("com.databricks.spark.csv").option("header","true").save(output2) + txt_count.repartition(1) + .write.format("com.databricks.spark.csv") + .option("header","true").save(output2) sc.stop()