Skip to content
Projects
Groups
Snippets
Help
Loading...
Sign in
Toggle navigation
F
ffm-baseline
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
ML
ffm-baseline
Commits
12ab4211
Commit
12ab4211
authored
Mar 19, 2020
by
赵威
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
update key
parent
7b103f10
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
8 additions
and
10 deletions
+8
-10
tag3_update_user_portrait_offline.py
eda/smart_rank/tag3_update_user_portrait_offline.py
+8
-10
No files found.
eda/smart_rank/tag3_update_user_portrait_offline.py
View file @
12ab4211
...
@@ -70,11 +70,14 @@ def update_tag3_user_portrait(cl_id, redis_client):
...
@@ -70,11 +70,14 @@ def update_tag3_user_portrait(cl_id, redis_client):
"projects"
:
projects_score
"projects"
:
projects_score
}
}
key
=
"doris:user_portrait:tag3:device_id:"
+
str
(
cl_id
)
key2
=
"doris:user_portrait:tag3:increment_update:device_id:"
+
str
(
cl_id
)
redis_client
=
redis
.
StrictRedis
.
from_url
(
"redis://:ReDis!GmTx*0aN9@172.16.40.173:6379"
)
if
(
len
(
first_demands_score
.
keys
())
>
0
)
or
(
len
(
second_demands_score
.
keys
())
>
0
)
or
\
if
(
len
(
first_demands_score
.
keys
())
>
0
)
or
(
len
(
second_demands_score
.
keys
())
>
0
)
or
\
(
len
(
first_solutions_score
.
keys
())
>
0
)
or
(
len
(
second_solutions_score
.
keys
())
>
0
)
or
\
(
len
(
first_solutions_score
.
keys
())
>
0
)
or
(
len
(
second_solutions_score
.
keys
())
>
0
)
or
\
(
len
(
first_positions_score
.
keys
())
>
0
)
or
(
len
(
second_positions_score
.
keys
())
>
0
)
or
\
(
len
(
first_positions_score
.
keys
())
>
0
)
or
(
len
(
second_positions_score
.
keys
())
>
0
)
or
\
(
len
(
projects_score
.
keys
())
>
0
):
(
len
(
projects_score
.
keys
())
>
0
):
key
=
"doris:user_portrait:tag3:device_id:"
+
str
(
cl_id
)
redis_client
.
set
(
key
,
json
.
dumps
(
res
))
redis_client
.
set
(
key
,
json
.
dumps
(
res
))
redis_client
.
expire
(
key
,
60
*
60
*
24
*
30
)
redis_client
.
expire
(
key
,
60
*
60
*
24
*
30
)
...
@@ -88,8 +91,6 @@ def update_tag3_user_portrait(cl_id, redis_client):
...
@@ -88,8 +91,6 @@ def update_tag3_user_portrait(cl_id, redis_client):
"second_positions"
:
list
(
second_positions_score
.
keys
()),
"second_positions"
:
list
(
second_positions_score
.
keys
()),
"projects"
:
list
(
projects_score
.
keys
())
"projects"
:
list
(
projects_score
.
keys
())
}
}
key2
=
"doris:user_portrait:tag3:increment_update:device_id:"
+
str
(
cl_id
)
redis_client
.
delete
(
key2
)
redis_client
.
set
(
key2
,
json
.
dumps
(
res2
))
redis_client
.
set
(
key2
,
json
.
dumps
(
res2
))
redis_client
.
expire
(
key2
,
60
*
60
*
24
*
30
)
redis_client
.
expire
(
key2
,
60
*
60
*
24
*
30
)
...
@@ -97,6 +98,9 @@ def update_tag3_user_portrait(cl_id, redis_client):
...
@@ -97,6 +98,9 @@ def update_tag3_user_portrait(cl_id, redis_client):
","
.
join
(
first_demands_score
.
keys
()),
","
.
join
(
second_demands_score
.
keys
()),
","
.
join
(
first_demands_score
.
keys
()),
","
.
join
(
second_demands_score
.
keys
()),
","
.
join
(
first_positions_score
.
keys
()),
","
.
join
(
second_positions_score
.
keys
()),
","
.
join
(
first_positions_score
.
keys
()),
","
.
join
(
second_positions_score
.
keys
()),
","
.
join
(
projects_score
.
keys
()))
","
.
join
(
projects_score
.
keys
()))
else
:
redis_client
.
delete
(
key
)
redis_client
.
delete
(
key2
)
return
cl_id
return
cl_id
...
@@ -124,10 +128,8 @@ def consume_kafka():
...
@@ -124,10 +128,8 @@ def consume_kafka():
spark
.
sparkContext
.
setLogLevel
(
"WARN"
)
spark
.
sparkContext
.
setLogLevel
(
"WARN"
)
spark
.
sparkContext
.
addPyFile
(
"/srv/apps/ffm-baseline_git/eda/smart_rank/tool.py"
)
spark
.
sparkContext
.
addPyFile
(
"/srv/apps/ffm-baseline_git/eda/smart_rank/tool.py"
)
redis_client
=
redis
.
StrictRedis
.
from_url
(
"redis://:ReDis!GmTx*0aN9@172.16.40.173:6379"
)
device_ids_lst_rdd
=
spark
.
sparkContext
.
parallelize
(
device_ids_lst
,
numSlices
=
1000
)
device_ids_lst_rdd
=
spark
.
sparkContext
.
parallelize
(
device_ids_lst
,
numSlices
=
1000
)
result
=
device_ids_lst_rdd
.
repartition
(
100
)
.
map
(
lambda
x
:
update_tag3_user_portrait
(
x
,
redis_client
))
result
=
device_ids_lst_rdd
.
repartition
(
100
)
.
map
(
lambda
x
:
update_tag3_user_portrait
(
x
))
# result.foreach(print)
# result.foreach(print)
result
.
collect
()
result
.
collect
()
spark
.
stop
()
spark
.
stop
()
...
@@ -136,10 +138,6 @@ def consume_kafka():
...
@@ -136,10 +138,6 @@ def consume_kafka():
if
__name__
==
"__main__"
:
if
__name__
==
"__main__"
:
# cl_id = "866017030837899"
# res = update_tag3_user_portrait(cl_id)
# print(res)
start
=
datetime
.
datetime
.
now
()
start
=
datetime
.
datetime
.
now
()
consume_kafka
()
consume_kafka
()
end
=
datetime
.
datetime
.
now
()
end
=
datetime
.
datetime
.
now
()
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment