Skip to content
Projects
Groups
Snippets
Help
Loading...
Sign in
Toggle navigation
C
crawler
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
backend
crawler
Commits
099cb699
Commit
099cb699
authored
4 years ago
by
litaolemo
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
update
parent
e6507b07
Hide whitespace changes
Inline
Side-by-side
Showing
1 changed file
with
3 additions
and
3 deletions
+3
-3
cal_ni_and_put_to_backend.py
crawler_sys/scheduler/cal_ni_and_put_to_backend.py
+3
-3
No files found.
crawler_sys/scheduler/cal_ni_and_put_to_backend.py
View file @
099cb699
...
...
@@ -153,11 +153,11 @@ def scan_from_redis(push_rule_class_list) -> Dict:
set_name
=
"exists_doc_id_set_
%
s"
%
datetime
.
datetime
.
fromtimestamp
(
one_data
[
"release_time"
]
/
1e3
)
.
strftime
(
"
%
Y-
%
m-
%
d"
)
if
rds
.
sismember
(
set_name
,
one_data
[
"doc_id"
]):
break
continue
else
:
rds
.
sadd
(
set_name
,
one_data
[
"doc_id"
])
yield
one_data
break
continue
# print(res_list)
...
...
@@ -464,8 +464,8 @@ def task_main():
# 循环处理抓取数据,返回需要添加至后端的数据
for
res_data
in
scan_from_redis
(
rules_list
):
# 符合规则入库数据
tractate_id
=
write_data_into_mysql
(
res_data
)
try
:
tractate_id
=
write_data_into_mysql
(
res_data
)
print
(
"line394"
,
tractate_id
,
res_data
)
if
res_data
[
"level"
]
>=
3
and
tractate_id
:
title_str
=
res_data
[
"platform"
]
+
"帖子内容审核"
...
...
This diff is collapsed.
Click to expand it.
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment