Commit c4ace8b2 authored by litaolemo's avatar litaolemo

update

parent 7545d0ea
...@@ -73,7 +73,7 @@ spark.sql("CREATE TEMPORARY FUNCTION is_json AS 'com.gmei.hive.common.udf.UDFJso ...@@ -73,7 +73,7 @@ spark.sql("CREATE TEMPORARY FUNCTION is_json AS 'com.gmei.hive.common.udf.UDFJso
spark.sql("CREATE TEMPORARY FUNCTION arrayMerge AS 'com.gmei.hive.common.udf.UDFArryMerge'") spark.sql("CREATE TEMPORARY FUNCTION arrayMerge AS 'com.gmei.hive.common.udf.UDFArryMerge'")
task_list = [] task_list = []
task_days = 3 task_days = 90
for t in range(0, task_days): for t in range(0, task_days):
day_num = 0 - t day_num = 0 - t
now = (datetime.datetime.now() + datetime.timedelta(days=day_num)) now = (datetime.datetime.now() + datetime.timedelta(days=day_num))
...@@ -166,7 +166,7 @@ group by t1.partition_date,device_os_type,active_type,channel ...@@ -166,7 +166,7 @@ group by t1.partition_date,device_os_type,active_type,channel
print(res) print(res)
device_os_type = res.device_os_type device_os_type = res.device_os_type
active_type = res.active_type active_type = res.active_type
partition_date = yesterday_str partition_date = today_str
channel = res.channel channel = res.channel
pid = hashlib.md5((partition_date + device_os_type + active_type + channel).encode("utf8")).hexdigest() pid = hashlib.md5((partition_date + device_os_type + active_type + channel).encode("utf8")).hexdigest()
diary_click_num = res.diary_click_pv diary_click_num = res.diary_click_pv
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment