Skip to content
Projects
Groups
Snippets
Help
Loading...
Sign in
Toggle navigation
C
crawler
Project
Project
Details
Activity
Cycle Analytics
Repository
Repository
Files
Commits
Branches
Tags
Contributors
Graph
Compare
Charts
Issues
0
Issues
0
List
Board
Labels
Milestones
Merge Requests
0
Merge Requests
0
CI / CD
CI / CD
Pipelines
Jobs
Schedules
Charts
Wiki
Wiki
Snippets
Snippets
Members
Members
Collapse sidebar
Close sidebar
Activity
Graph
Charts
Create a new issue
Jobs
Commits
Issue Boards
Open sidebar
backend
crawler
Commits
6e5c5a8f
Commit
6e5c5a8f
authored
4 years ago
by
litaolemo
Browse files
Options
Browse Files
Download
Email Patches
Plain Diff
update
parent
0fe0b5ab
No related merge requests found
Hide whitespace changes
Inline
Side-by-side
Showing
2 changed files
with
3 additions
and
2 deletions
+3
-2
update_data_in_target_releasers_multi_process_by_date_from_redis.py
...a_in_target_releasers_multi_process_by_date_from_redis.py
+1
-1
crawler_weibo.py
crawler_sys/site_crawler_by_redis/crawler_weibo.py
+2
-1
No files found.
crawler_sys/framework/update_data_in_target_releasers_multi_process_by_date_from_redis.py
View file @
6e5c5a8f
...
...
@@ -71,7 +71,7 @@ parser.add_argument('-name', '--name', default="crawler01", type=str,
help
=
(
'this computer name'
))
args
=
parser
.
parse_args
()
rds_save
=
redis
.
StrictRedis
(
host
=
'172.1
6.40.164
'
,
port
=
6379
,
db
=
20
,
decode_responses
=
True
,
password
=
'ReDis!GmTx*0aN12'
)
rds_save
=
redis
.
StrictRedis
(
host
=
'172.1
8.51.10
'
,
port
=
6379
,
db
=
20
,
decode_responses
=
True
,
password
=
'ReDis!GmTx*0aN12'
)
rds_1
=
redis
.
StrictRedis
(
host
=
'172.18.51.10'
,
port
=
6379
,
db
=
19
,
decode_responses
=
True
)
rds_17
=
redis
.
StrictRedis
(
host
=
'172.18.51.10'
,
port
=
6379
,
db
=
17
,
decode_responses
=
True
)
platform_crawler_reg
=
{
...
...
This diff is collapsed.
Click to expand it.
crawler_sys/site_crawler_by_redis/crawler_weibo.py
View file @
6e5c5a8f
...
...
@@ -127,6 +127,7 @@ class Crawler_weibo():
except
:
get_page
=
None
has_more
=
False
print
(
get_page
.
text
)
if
get_page
and
get_page
.
status_code
==
200
:
try
:
page_json
=
get_page
.
json
()
...
...
@@ -330,7 +331,7 @@ if __name__ == '__main__':
]
import
redis
rds
=
redis
.
StrictRedis
(
host
=
'172.18.51.10'
,
port
=
6379
,
db
=
17
,
decode_responses
=
True
)
rds
=
redis
.
StrictRedis
(
host
=
'172.18.51.10'
,
port
=
6379
,
db
=
20
,
decode_responses
=
True
)
for
url
in
url_list
:
res
=
test
.
releaser_page
(
url
,
output_to_es_raw
=
True
,
es_index
=
'crawler-data-raw'
,
...
...
This diff is collapsed.
Click to expand it.
Write
Preview
Markdown
is supported
0%
Try again
or
attach a new file
Attach a file
Cancel
You are about to add
0
people
to the discussion. Proceed with caution.
Finish editing this message first!
Cancel
Please
register
or
sign in
to comment