diff --git a/eda/feededa/src/main/scala/com/gmei/GmeiConfig.scala b/eda/feededa/src/main/scala/com/gmei/GmeiConfig.scala
index 0c62473d33685fe44872a8a30a1a3547f850fe16..830795d3f7ab892632a68ab435cac71522a35c8c 100644
--- a/eda/feededa/src/main/scala/com/gmei/GmeiConfig.scala
+++ b/eda/feededa/src/main/scala/com/gmei/GmeiConfig.scala
@@ -48,6 +48,7 @@ object GmeiConfig extends Serializable {
       .builder()
       .config(sparkConf)
       .appName("feededa")
+      .enableHiveSupport()
       .getOrCreate()
 
     val context = SparkContext.getOrCreate(sparkConf)
diff --git a/eda/feededa/src/main/scala/com/gmei/WeafareStat.scala b/eda/feededa/src/main/scala/com/gmei/WeafareStat.scala
index bb6b8bedcb5fc7849208ee043d7e2fc88f6aa242..87e0507cc62945a2b9f0d012612730cb98a5a6f4 100644
--- a/eda/feededa/src/main/scala/com/gmei/WeafareStat.scala
+++ b/eda/feededa/src/main/scala/com/gmei/WeafareStat.scala
@@ -81,7 +81,9 @@ object WeafareStat {
       )
       video_count.show()
       val output1 = "./test_vd_cids.csv"
-      video_count.repartition(1).write.format("com.databricks.spark.csv").option("header","true").save(output1)
+      video_count.repartition(1)
+        .write.format("com.databricks.spark.csv")
+        .option("header","true").save(output1)
 
       val txt_count = sc.sql(
         s"""
@@ -95,7 +97,9 @@ object WeafareStat {
       )
       txt_count.show()
       val output2 = "./test_txt_cids.csv"
-      txt_count.repartition(1).write.format("com.databricks.spark.csv").option("header","true").save(output2)
+      txt_count.repartition(1)
+        .write.format("com.databricks.spark.csv")
+        .option("header","true").save(output2)
 
 
       sc.stop()