Commit 80fbb852 authored by 王志伟's avatar 王志伟

更改时间

parent 1623a6dd
...@@ -8,11 +8,11 @@ import org.apache.spark.sql.SaveMode ...@@ -8,11 +8,11 @@ import org.apache.spark.sql.SaveMode
import org.apache.log4j.{Level, Logger} import org.apache.log4j.{Level, Logger}
import scopt.OptionParser import scopt.OptionParser
import com.gmei.lib.AbstractParams import com.gmei.lib.AbstractParams
//import com.sun.deploy.uitoolkit.Window
import org.apache.spark.sql.expressions.Window import org.apache.spark.sql.expressions.Window
import org.apache.spark.sql.functions.row_number import org.apache.spark.sql.functions.row_number
object data_feed_exposure_precise { object data_feed_exposure_precise {
Logger.getLogger("org.apache.spark").setLevel(Level.WARN) Logger.getLogger("org.apache.spark").setLevel(Level.WARN)
...@@ -654,6 +654,7 @@ object tag_value { ...@@ -654,6 +654,7 @@ object tag_value {
val sc = spark_env._2 val sc = spark_env._2
sc.sql("use jerry_prod") sc.sql("use jerry_prod")
import sc.implicits._
val stat_date = GmeiConfig.getMinusNDate(1) val stat_date = GmeiConfig.getMinusNDate(1)
...@@ -759,6 +760,8 @@ object tag_value { ...@@ -759,6 +760,8 @@ object tag_value {
val result = temp2.select("tag_level").distinct().withColumn("index", row_number().over(w)) val result = temp2.select("tag_level").distinct().withColumn("index", row_number().over(w))
result.show() result.show()
val test=result.rdd.map(row=>(row(0).toString,row(1).toString)).map(row=>(row._1.split(",").head,row._1.split(",")(1),row._2)).toDF().show()
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment