Skip to content
Projects
Groups
Snippets
Help
Loading...
Sign in
Toggle navigation
F
ffm-baseline
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
ML
ffm-baseline
Commits
abb9f06b
Commit
abb9f06b
authored
Sep 27, 2019
by
张彦钊
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
add
parent
772f5536
Show whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
10 additions
and
8 deletions
+10
-8
hello.py
hello.py
+10
-8
No files found.
hello.py
View file @
abb9f06b
...
@@ -11,6 +11,7 @@ from pyspark import StorageLevel
...
@@ -11,6 +11,7 @@ from pyspark import StorageLevel
def
position
(
x
,
y
):
def
position
(
x
,
y
):
sum
=
0
sum
=
0
for
i
in
x
:
for
i
in
x
:
if
'absolute_position'
in
i
:
if
int
(
i
[
'absolute_position'
])
<=
y
:
if
int
(
i
[
'absolute_position'
])
<=
y
:
sum
=
sum
+
1
sum
=
sum
+
1
return
sum
return
sum
...
@@ -18,6 +19,7 @@ def position(x,y):
...
@@ -18,6 +19,7 @@ def position(x,y):
def
cpc
(
x
,
y
):
def
cpc
(
x
,
y
):
sum
=
0
sum
=
0
for
i
in
x
:
for
i
in
x
:
if
(
"is_cpc"
in
i
)
and
(
'absolute_position'
in
i
):
if
(
int
(
i
[
'absolute_position'
])
<=
y
)
and
(
i
[
"is_cpc"
]
==
1
):
if
(
int
(
i
[
'absolute_position'
])
<=
y
)
and
(
i
[
"is_cpc"
]
==
1
):
sum
=
sum
+
1
sum
=
sum
+
1
return
sum
return
sum
...
@@ -37,34 +39,34 @@ if __name__ == '__main__':
...
@@ -37,34 +39,34 @@ if __name__ == '__main__':
sql
=
"select params['exposure_cards'] from online.ml_community_precise_exposure_detail "
\
sql
=
"select params['exposure_cards'] from online.ml_community_precise_exposure_detail "
\
"where action = 'page_precise_exposure' and page_name = 'search_result_welfare' "
\
"where action = 'page_precise_exposure' and page_name = 'search_result_welfare' "
\
"AND partition_date='20190926'"
"AND partition_date='20190926'"
df
=
spark
.
sql
(
sql
)
df
=
spark
.
sql
(
sql
)
.
rdd
# # df.show(6)
# # df.show(6)
# # params['exposure_cards'],
# # params['exposure_cards'],
df
.
persist
()
df
.
persist
()
total
=
[]
total
=
[]
rdd
=
df
.
rdd
.
map
(
lambda
x
:(
"a"
,
position
(
eval
(
x
[
0
]),
10
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
rdd
=
df
.
map
(
lambda
x
:(
"a"
,
position
(
eval
(
x
[
0
]),
10
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
tmp
=
rdd
.
collect
()[
0
]
tmp
=
rdd
.
collect
()[
0
]
total
.
append
(
tmp
)
total
.
append
(
tmp
)
rdd
=
df
.
rdd
.
map
(
lambda
x
:
(
"a"
,
position
(
eval
(
x
[
0
]),
30
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
rdd
=
df
.
map
(
lambda
x
:
(
"a"
,
position
(
eval
(
x
[
0
]),
30
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
tmp
=
rdd
.
collect
()[
0
]
tmp
=
rdd
.
collect
()[
0
]
total
.
append
(
tmp
)
total
.
append
(
tmp
)
rdd
=
df
.
rdd
.
map
(
lambda
x
:
(
"a"
,
position
(
eval
(
x
[
0
]),
50
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
rdd
=
df
.
map
(
lambda
x
:
(
"a"
,
position
(
eval
(
x
[
0
]),
50
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
tmp
=
rdd
.
collect
()[
0
]
tmp
=
rdd
.
collect
()[
0
]
total
.
append
(
tmp
)
total
.
append
(
tmp
)
rdd
=
df
.
rdd
.
map
(
lambda
x
:
(
"a"
,
cpc
(
eval
(
x
[
0
]),
10
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
rdd
=
df
.
map
(
lambda
x
:
(
"a"
,
cpc
(
eval
(
x
[
0
]),
10
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
tmp
=
rdd
.
collect
()[
0
]
tmp
=
rdd
.
collect
()[
0
]
total
.
append
(
tmp
)
total
.
append
(
tmp
)
rdd
=
df
.
rdd
.
map
(
lambda
x
:
(
"a"
,
cpc
(
eval
(
x
[
0
]),
30
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
rdd
=
df
.
map
(
lambda
x
:
(
"a"
,
cpc
(
eval
(
x
[
0
]),
30
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
tmp
=
rdd
.
collect
()[
0
]
tmp
=
rdd
.
collect
()[
0
]
total
.
append
(
tmp
)
total
.
append
(
tmp
)
rdd
=
df
.
rdd
.
map
(
lambda
x
:
(
"a"
,
cpc
(
eval
(
x
[
0
]),
50
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
rdd
=
df
.
map
(
lambda
x
:
(
"a"
,
cpc
(
eval
(
x
[
0
]),
50
)))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
tmp
=
rdd
.
collect
()[
0
]
tmp
=
rdd
.
collect
()[
0
]
total
.
append
(
tmp
)
total
.
append
(
tmp
)
rdd
=
df
.
rdd
.
map
(
lambda
x
:
(
"a"
,
len
(
eval
(
x
[
0
]))))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
rdd
=
df
.
map
(
lambda
x
:
(
"a"
,
len
(
eval
(
x
[
0
]))))
.
reduceByKey
(
lambda
x
,
y
:
x
+
y
)
.
map
(
lambda
x
:
x
[
1
])
tmp
=
rdd
.
collect
()[
0
]
tmp
=
rdd
.
collect
()[
0
]
total
.
append
(
tmp
)
total
.
append
(
tmp
)
...
...
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment