Commit 337978fc authored by 赵威's avatar 赵威

try sort res

parent adf38dfe
......@@ -5,6 +5,7 @@ import operator
from collections import Counter
import redis
from pyspark import SparkConf
from pyspark.sql import SparkSession
from tool import (get_jerry_test, get_tag3_user_log, send_email, write_user_portrait)
......@@ -13,7 +14,8 @@ from tool import (get_jerry_test, get_tag3_user_log, send_email, write_user_port
# [{'激光': 1.949194898204873}, {'手术': 1.949194898204873}, {'手术': 1.949194898204873}, {'手术': 1.949194898204873}]
# {'手术': 5.8475846946146195, '激光': 1.949194898204873}
def merge_values(list_of_dict):
return dict(functools.reduce(operator.add, map(Counter, list_of_dict)))
d = dict(functools.reduce(operator.add, map(Counter, list_of_dict)))
return dict(sorted(d.items(), key=lambda x: x[1], reverse=True))
# [("a", 1), ("b", 2)]
......@@ -69,16 +71,18 @@ def update_tag3_user_portrait(cl_id):
"second_positions": second_positions_score,
"projects": projects_score
}
print(cl_id)
print(res)
key = "doris:user_portrait:tag3:device_id:" + str(cl_id)
redis_client = redis.StrictRedis.from_url("redis://:ReDis!GmTx*0aN9@172.16.40.173:6379")
if (len(first_demands_score.keys()) > 0) or (len(second_demands_score.keys()) > 0) or \
(len(first_solutions_score.keys()) > 0) or (len(second_solutions_score.keys()) > 0) or \
(len(first_positions_score.keys()) > 0) or (len(second_positions_score.keys()) > 0) or \
(len(projects_score.keys()) > 0):
redis_client.set(key, json.dumps(res))
redis_client.expire(key, 60 * 60 * 24 * 30)
# if (len(first_demands_score.keys()) > 0) or (len(second_demands_score.keys()) > 0) or \
# (len(first_solutions_score.keys()) > 0) or (len(second_solutions_score.keys()) > 0) or \
# (len(first_positions_score.keys()) > 0) or (len(second_positions_score.keys()) > 0) or \
# (len(projects_score.keys()) > 0):
# redis_client.set(key, json.dumps(res))
# redis_client.expire(key, 60 * 60 * 24 * 30)
# only need the first time
# key2 = "doris:user_portrait:tag3:increment_update:device_id:" + str(cl_id)
......@@ -94,12 +98,12 @@ def update_tag3_user_portrait(cl_id):
# redis_client.set(key2, json.dumps(res2))
# redis_client.expire(key2, 60 * 60 * 24 * 30)
write_user_portrait(cl_id, ",".join(first_solutions_score.keys()), ",".join(second_solutions_score.keys()),
",".join(first_demands_score.keys()), ",".join(second_demands_score.keys()),
",".join(first_positions_score.keys()), ",".join(second_positions_score.keys()),
",".join(projects_score.keys()))
# write_user_portrait(cl_id, ",".join(first_solutions_score.keys()), ",".join(second_solutions_score.keys()),
# ",".join(first_demands_score.keys()), ",".join(second_demands_score.keys()),
# ",".join(first_positions_score.keys()), ",".join(second_positions_score.keys()),
# ",".join(projects_score.keys()))
return cl_id
return cl_id + str(res)
def consume_kafka():
......
Markdown is supported
0% or
You are about to add 0 people to the discussion. Proceed with caution.
Finish editing this message first!
Please register or to comment