SpiderClub / haipproxy

:sparkling_heart: High available distributed ip proxy pool, powerd by Scrapy and Redis
https://spiderclub.github.io/haipproxy/
MIT License
5.43k stars 914 forks source link

win10 成功安装,但是redis中只有init_queue,其他的validated什么都都没有 #86

Open lmst2 opened 5 years ago

lmst2 commented 5 years ago

wechat screenshot_20181129204155

win10 成功安装,但是redis中只有init_queue,其他的validated什么的都没有 haipproxy本体是在win10 pro的电脑上,redis装在树莓派里,但是既然其他的queue能创建,应该不是连接的关系。 能看一下是哪里的问题吗

crawler输出:

2018-11-29 19:51:31 [validator] INFO: crawler scheduler is starting... 2018-11-29 19:51:36 [validator] INFO: crawler task mogumiao has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task xdaili has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task xici has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task kuaidaili has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task nianshao has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task kxdaili has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task 66ip has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task ip181 has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task mrhinkydink has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task coderbusy has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task baizhongsou has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task ip3366 has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task swei360 has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task data5u has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task proxylists has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task iphai has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task yundaili has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task httpsdaili has been stored into redis successfully 2018-11-29 19:51:36 [validator] INFO: crawler task my-proxy has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task atomintersoft has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task ab57 has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task cool-proxy has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task us-proxy has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task rmccurdy has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task xroxy has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task goubanjia has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task free-proxy has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task proxylistplus has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task proxydb has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task proxy-list has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task cn-proxy has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task cnproxy has been stored into redis successfully 2018-11-29 19:51:37 [validator] INFO: crawler task gatherproxy has been stored into redis successfully 2018-11-29 19:56:37 [schedule] INFO: Running job Every 5 minutes do schedule_task_with_lock({'name': 'mogumiao', 'resource': ['http://www.mogumiao.com/proxy/free/listFreeIp', 'http://www.mogumiao.com/proxy/api/freeIp?count=15'], 'task_queue': 'haipproxy:spider:common', 'internal': 5, 'enable': 1}) (last run: [never], next run: 2018-11-29 19:56:37) 2018-11-29 19:56:38 [validator] INFO: crawler task mogumiao has been stored into redis successfully 2018-11-29 20:01:38 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'xdaili', 'resource': ['http://www.xdaili.cn:80/ipagent/freeip/getFreeIps?page=1&rows=10'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:01:37) 2018-11-29 20:01:38 [validator] INFO: crawler task xdaili has been stored into redis successfully 2018-11-29 20:01:38 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'data5u', 'resource': ['http://www.data5u.com/free/index.shtml', 'http://www.data5u.com/free/gngn/index.shtml', 'http://www.data5u.com/free/gwgn/index.shtml'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:01:37) 2018-11-29 20:01:38 [validator] INFO: crawler task data5u has been stored into redis successfully 2018-11-29 20:01:38 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'ip181', 'resource': ['http://www.ip181.com/', 'http://www.ip181.com/daili/1.html', 'http://www.ip181.com/daili/2.html', 'http://www.ip181.com/daili/3.html'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:01:37) 2018-11-29 20:01:38 [validator] INFO: crawler task ip181 has been stored into redis successfully 2018-11-29 20:01:38 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'goubanjia', 'resource': ['http://www.goubanjia.com/'], 'task_queue': 'haipproxy:spider:ajax', 'internal': 10, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:01:37) 2018-11-29 20:01:38 [validator] INFO: crawler task goubanjia has been stored into redis successfully 2018-11-29 20:01:38 [schedule] INFO: Running job Every 5 minutes do schedule_task_with_lock({'name': 'mogumiao', 'resource': ['http://www.mogumiao.com/proxy/free/listFreeIp', 'http://www.mogumiao.com/proxy/api/freeIp?count=15'], 'task_queue': 'haipproxy:spider:common', 'internal': 5, 'enable': 1}) (last run: 2018-11-29 19:56:38, next run: 2018-11-29 20:01:38) 2018-11-29 20:01:38 [validator] INFO: crawler task mogumiao has been stored into redis successfully 2018-11-29 20:06:38 [schedule] INFO: Running job Every 5 minutes do schedule_task_with_lock({'name': 'mogumiao', 'resource': ['http://www.mogumiao.com/proxy/free/listFreeIp', 'http://www.mogumiao.com/proxy/api/freeIp?count=15'], 'task_queue': 'haipproxy:spider:common', 'internal': 5, 'enable': 1}) (last run: 2018-11-29 20:01:38, next run: 2018-11-29 20:06:38) 2018-11-29 20:06:38 [validator] INFO: crawler task mogumiao has been stored into redis successfully 2018-11-29 20:11:39 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'xdaili', 'resource': ['http://www.xdaili.cn:80/ipagent/freeip/getFreeIps?page=1&rows=10'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:01:38, next run: 2018-11-29 20:11:38) 2018-11-29 20:11:39 [validator] INFO: crawler task xdaili has been stored into redis successfully 2018-11-29 20:11:39 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'data5u', 'resource': ['http://www.data5u.com/free/index.shtml', 'http://www.data5u.com/free/gngn/index.shtml', 'http://www.data5u.com/free/gwgn/index.shtml'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:01:38, next run: 2018-11-29 20:11:38) 2018-11-29 20:11:39 [validator] INFO: crawler task data5u has been stored into redis successfully 2018-11-29 20:11:39 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'ip181', 'resource': ['http://www.ip181.com/', 'http://www.ip181.com/daili/1.html', 'http://www.ip181.com/daili/2.html', 'http://www.ip181.com/daili/3.html'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:01:38, next run: 2018-11-29 20:11:38) 2018-11-29 20:11:39 [validator] INFO: crawler task ip181 has been stored into redis successfully 2018-11-29 20:11:39 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'goubanjia', 'resource': ['http://www.goubanjia.com/'], 'task_queue': 'haipproxy:spider:ajax', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:01:38, next run: 2018-11-29 20:11:38) 2018-11-29 20:11:39 [validator] INFO: crawler task goubanjia has been stored into redis successfully 2018-11-29 20:11:39 [schedule] INFO: Running job Every 5 minutes do schedule_task_with_lock({'name': 'mogumiao', 'resource': ['http://www.mogumiao.com/proxy/free/listFreeIp', 'http://www.mogumiao.com/proxy/api/freeIp?count=15'], 'task_queue': 'haipproxy:spider:common', 'internal': 5, 'enable': 1}) (last run: 2018-11-29 20:06:38, next run: 2018-11-29 20:11:38) 2018-11-29 20:11:41 [validator] INFO: crawler task mogumiao has been stored into redis successfully 2018-11-29 20:16:41 [schedule] INFO: Running job Every 5 minutes do schedule_task_with_lock({'name': 'mogumiao', 'resource': ['http://www.mogumiao.com/proxy/free/listFreeIp', 'http://www.mogumiao.com/proxy/api/freeIp?count=15'], 'task_queue': 'haipproxy:spider:common', 'internal': 5, 'enable': 1}) (last run: 2018-11-29 20:11:41, next run: 2018-11-29 20:16:41) 2018-11-29 20:16:41 [validator] INFO: crawler task mogumiao has been stored into redis successfully 2018-11-29 20:21:37 [schedule] INFO: Running job Every 30 minutes do schedule_task_with_lock({'name': 'baizhongsou', 'resource': ['http://ip.baizhongsou.com/'], 'task_queue': 'haipproxy:spider:common', 'internal': 30, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:21:37) 2018-11-29 20:21:37 [validator] INFO: crawler task baizhongsou has been stored into redis successfully 2018-11-29 20:21:37 [schedule] INFO: Running job Every 30 minutes do schedule_task_with_lock({'name': 'ip3366', 'resource': ['http://www.ip3366.net/free/?stype=1&page=1', 'http://www.ip3366.net/free/?stype=1&page=2', 'http://www.ip3366.net/free/?stype=3&page=1', 'http://www.ip3366.net/free/?stype=3&page=2'], 'task_queue': 'haipproxy:spider:common', 'internal': 30, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:21:37) 2018-11-29 20:21:37 [validator] INFO: crawler task ip3366 has been stored into redis successfully 2018-11-29 20:21:38 [schedule] INFO: Running job Every 30 minutes do schedule_task_with_lock({'name': 'swei360', 'resource': ['http://www.swei360.com/free/?page=1', 'http://www.swei360.com/free/?page=2', 'http://www.swei360.com/free/?page=3', 'http://www.swei360.com/free/?stype=3&page=1', 'http://www.swei360.com/free/?stype=3&page=2', 'http://www.swei360.com/free/?stype=3&page=3'], 'task_queue': 'haipproxy:spider:common', 'internal': 30, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:21:37) 2018-11-29 20:21:38 [validator] INFO: crawler task swei360 has been stored into redis successfully 2018-11-29 20:21:38 [schedule] INFO: Running job Every 30 minutes do schedule_task_with_lock({'name': 'cool-proxy', 'resource': ['https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:1', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:2', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:3', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:4', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:5', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:6', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:7', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:8', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:9', 'https://www.cool-proxy.net/proxies/http_proxy_list/country_code:/port:/anonymous:1/page:10'], 'task_queue': 'haipproxy:spider:ajax', 'internal': 30, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:21:37) 2018-11-29 20:21:38 [validator] INFO: crawler task cool-proxy has been stored into redis successfully 2018-11-29 20:21:40 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'xdaili', 'resource': ['http://www.xdaili.cn:80/ipagent/freeip/getFreeIps?page=1&rows=10'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:11:39, next run: 2018-11-29 20:21:39) 2018-11-29 20:21:40 [validator] INFO: crawler task xdaili has been stored into redis successfully 2018-11-29 20:21:40 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'data5u', 'resource': ['http://www.data5u.com/free/index.shtml', 'http://www.data5u.com/free/gngn/index.shtml', 'http://www.data5u.com/free/gwgn/index.shtml'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:11:39, next run: 2018-11-29 20:21:39) 2018-11-29 20:21:40 [validator] INFO: crawler task data5u has been stored into redis successfully 2018-11-29 20:21:40 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'ip181', 'resource': ['http://www.ip181.com/', 'http://www.ip181.com/daili/1.html', 'http://www.ip181.com/daili/2.html', 'http://www.ip181.com/daili/3.html'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:11:39, next run: 2018-11-29 20:21:39) 2018-11-29 20:21:40 [validator] INFO: crawler task ip181 has been stored into redis successfully 2018-11-29 20:21:40 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'goubanjia', 'resource': ['http://www.goubanjia.com/'], 'task_queue': 'haipproxy:spider:ajax', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:11:39, next run: 2018-11-29 20:21:39) 2018-11-29 20:21:40 [validator] INFO: crawler task goubanjia has been stored into redis successfully 2018-11-29 20:21:42 [schedule] INFO: Running job Every 5 minutes do schedule_task_with_lock({'name': 'mogumiao', 'resource': ['http://www.mogumiao.com/proxy/free/listFreeIp', 'http://www.mogumiao.com/proxy/api/freeIp?count=15'], 'task_queue': 'haipproxy:spider:common', 'internal': 5, 'enable': 1}) (last run: 2018-11-29 20:16:41, next run: 2018-11-29 20:21:41) 2018-11-29 20:21:42 [validator] INFO: crawler task mogumiao has been stored into redis successfully 2018-11-29 20:26:42 [schedule] INFO: Running job Every 5 minutes do schedule_task_with_lock({'name': 'mogumiao', 'resource': ['http://www.mogumiao.com/proxy/free/listFreeIp', 'http://www.mogumiao.com/proxy/api/freeIp?count=15'], 'task_queue': 'haipproxy:spider:common', 'internal': 5, 'enable': 1}) (last run: 2018-11-29 20:21:42, next run: 2018-11-29 20:26:42) 2018-11-29 20:26:42 [validator] INFO: crawler task mogumiao has been stored into redis successfully 2018-11-29 20:31:40 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'xdaili', 'resource': ['http://www.xdaili.cn:80/ipagent/freeip/getFreeIps?page=1&rows=10'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:21:40, next run: 2018-11-29 20:31:40) 2018-11-29 20:31:40 [validator] INFO: crawler task xdaili has been stored into redis successfully 2018-11-29 20:31:40 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'data5u', 'resource': ['http://www.data5u.com/free/index.shtml', 'http://www.data5u.com/free/gngn/index.shtml', 'http://www.data5u.com/free/gwgn/index.shtml'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:21:40, next run: 2018-11-29 20:31:40) 2018-11-29 20:31:41 [validator] INFO: crawler task data5u has been stored into redis successfully 2018-11-29 20:31:41 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'ip181', 'resource': ['http://www.ip181.com/', 'http://www.ip181.com/daili/1.html', 'http://www.ip181.com/daili/2.html', 'http://www.ip181.com/daili/3.html'], 'task_queue': 'haipproxy:spider:common', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:21:40, next run: 2018-11-29 20:31:40) 2018-11-29 20:31:41 [validator] INFO: crawler task ip181 has been stored into redis successfully 2018-11-29 20:31:41 [schedule] INFO: Running job Every 10 minutes do schedule_task_with_lock({'name': 'goubanjia', 'resource': ['http://www.goubanjia.com/'], 'task_queue': 'haipproxy:spider:ajax', 'internal': 10, 'enable': 1}) (last run: 2018-11-29 20:21:40, next run: 2018-11-29 20:31:40) 2018-11-29 20:31:41 [validator] INFO: crawler task goubanjia has been stored into redis successfully 2018-11-29 20:31:43 [schedule] INFO: Running job Every 5 minutes do schedule_task_with_lock({'name': 'mogumiao', 'resource': ['http://www.mogumiao.com/proxy/free/listFreeIp', 'http://www.mogumiao.com/proxy/api/freeIp?count=15'], 'task_queue': 'haipproxy:spider:common', 'internal': 5, 'enable': 1}) (last run: 2018-11-29 20:26:42, next run: 2018-11-29 20:31:42) 2018-11-29 20:31:43 [validator] INFO: crawler task mogumiao has been stored into redis successfully

validator输出:

2018-11-29 19:51:36 [validator] INFO: validator scheduler is starting... 2018-11-29 19:51:41 [validator] WARNING: fetched no proxies from task http 2018-11-29 19:51:41 [validator] WARNING: fetched no proxies from task https 2018-11-29 20:11:42 [schedule] INFO: Running job Every 20 minutes do schedule_task_with_lock({'name': 'http', 'task_queue': 'haipproxy:http:temp', 'resource': 'haipproxy:validated:http', 'internal': 20, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:11:41) 2018-11-29 20:11:42 [validator] WARNING: fetched no proxies from task http 2018-11-29 20:11:42 [schedule] INFO: Running job Every 20 minutes do schedule_task_with_lock({'name': 'https', 'task_queue': 'haipproxy:https:temp', 'resource': 'haipproxy:validated:https', 'internal': 20, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:11:41) 2018-11-29 20:11:42 [validator] WARNING: fetched no proxies from task https 2018-11-29 20:11:43 [schedule] INFO: Running job Every 20 minutes do schedule_task_with_lock({'name': 'weibo', 'task_queue': 'haipproxy:weibo:temp', 'resource': 'haipproxy:validated:weibo', 'internal': 20, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:11:41) 2018-11-29 20:11:43 [schedule] INFO: Running job Every 20 minutes do schedule_task_with_lock({'name': 'zhihu', 'task_queue': 'haipproxy:zhihu:temp', 'resource': 'haipproxy:validated:zhihu', 'internal': 20, 'enable': 1}) (last run: [never], next run: 2018-11-29 20:11:41)

haipproxy log:

2018-11-29 19:51:23 [urllib3.connectionpool] DEBUG: Starting new HTTPS connection (1): httpbin.org 2018-11-29 19:51:23 [urllib3.connectionpool] DEBUG: https://httpbin.org:443 "GET /ip HTTP/1.1" 200 31 2018-11-29 19:51:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:27 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:27 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:27 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:27 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:27 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:27 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:28 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:28 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:28 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:29 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:51:29 [cralwer] INFO: Read 0 requests from haipproxy:spider:common 2018-11-29 19:51:29 [cralwer] INFO: Read 0 requests from haipproxy:spider:common 2018-11-29 19:51:29 [cralwer] INFO: Read 0 requests from haipproxy:spider:ajax 2018-11-29 19:51:29 [cralwer] INFO: Read 0 requests from haipproxy:spider:ajax 2018-11-29 19:51:29 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:51:29 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:51:29 [cralwer] INFO: Read 0 requests from haipproxy:spider:ajax_gfw 2018-11-29 19:51:29 [cralwer] INFO: Read 0 requests from haipproxy:spider:ajax_gfw 2018-11-29 19:51:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:33 [cralwer] INFO: Read 0 requests from haipproxy:spider:common 2018-11-29 19:51:33 [cralwer] INFO: Read 0 requests from haipproxy:spider:ajax 2018-11-29 19:51:34 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:51:34 [cralwer] INFO: Read 0 requests from haipproxy:spider:ajax_gfw 2018-11-29 19:51:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:39 [cralwer] INFO: Read 10 requests from haipproxy:spider:common 2018-11-29 19:51:40 [cralwer] INFO: Read 10 requests from haipproxy:spider:ajax 2018-11-29 19:51:40 [cralwer] INFO: Read 10 requests from haipproxy:spider:gfw 2018-11-29 19:51:40 [cralwer] INFO: Read 10 requests from haipproxy:spider:ajax_gfw 2018-11-29 19:51:40 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/country/?c=China>: HTTP status code is not handled or not allowed 2018-11-29 19:51:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:44 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/country/?c=Brazil>: HTTP status code is not handled or not allowed 2018-11-29 19:51:46 [cralwer] INFO: Read 10 requests from haipproxy:spider:common 2018-11-29 19:51:48 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/country/?c=Indonesia>: HTTP status code is not handled or not allowed 2018-11-29 19:51:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:52 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/country/?c=Russia>: HTTP status code is not handled or not allowed 2018-11-29 19:51:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:56 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/country/?c=United%20States>: HTTP status code is not handled or not allowed 2018-11-29 19:51:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:51:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:51:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:51:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:51:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:51:59 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/country/?c=Thailand>: HTTP status code is not handled or not allowed 2018-11-29 19:52:01 [cralwer] INFO: Read 10 requests from haipproxy:spider:gfw 2018-11-29 19:52:02 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/port/8080>: HTTP status code is not handled or not allowed 2018-11-29 19:52:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:06 [cralwer] INFO: Read 10 requests from haipproxy:spider:common 2018-11-29 19:52:07 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/port/3128>: HTTP status code is not handled or not allowed 2018-11-29 19:52:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:09 [cralwer] INFO: Read 10 requests from haipproxy:spider:ajax 2018-11-29 19:52:11 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/port/80>: HTTP status code is not handled or not allowed 2018-11-29 19:52:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:14 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/port/8118>: HTTP status code is not handled or not allowed 2018-11-29 19:52:14 [cralwer] INFO: Read 10 requests from haipproxy:spider:ajax_gfw 2018-11-29 19:52:14 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy6.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:17 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/>: HTTP status code is not handled or not allowed 2018-11-29 19:52:18 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:18 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:18 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:18 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:18 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:18 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy7.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:21 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/anonymity/?t=Elite>: HTTP status code is not handled or not allowed 2018-11-29 19:52:23 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy8.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:52:23 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:52:23 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:52:23 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:52:23 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:23 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:52:23 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:24 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.gatherproxy.com/proxylist/anonymity/?t=Anonymous>: HTTP status code is not handled or not allowed 2018-11-29 19:52:27 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy9.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:28 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:28 [scrapy.extensions.logstats] INFO: Crawled 26 pages (at 26 pages/min), scraped 959 items (at 959 items/min) 2018-11-29 19:52:28 [scrapy.extensions.logstats] INFO: Crawled 16 pages (at 16 pages/min), scraped 13 items (at 13 items/min) 2018-11-29 19:52:28 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:52:29 [scrapy.extensions.logstats] INFO: Crawled 17 pages (at 17 pages/min), scraped 0 items (at 0 items/min) 2018-11-29 19:52:30 [cralwer] INFO: Read 10 requests from haipproxy:spider:common 2018-11-29 19:52:31 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy10.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:33 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:35 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxyedu1.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:35 [cralwer] INFO: Read 4 requests from haipproxy:spider:gfw 2018-11-29 19:52:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:38 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:38 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxyedu2.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:38 [cralwer] INFO: Read 10 requests from haipproxy:spider:ajax_gfw 2018-11-29 19:52:39 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=6>: HTTP status code is not handled or not allowed 2018-11-29 19:52:42 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy1.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:42 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=7>: HTTP status code is not handled or not allowed 2018-11-29 19:52:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:43 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:45 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=8>: HTTP status code is not handled or not allowed 2018-11-29 19:52:45 [cralwer] INFO: Read 10 requests from haipproxy:spider:ajax 2018-11-29 19:52:47 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy2.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:48 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=9>: HTTP status code is not handled or not allowed 2018-11-29 19:52:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:48 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:51 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy3.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:51 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=10>: HTTP status code is not handled or not allowed 2018-11-29 19:52:52 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:52:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:53 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:53 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:52:55 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy4.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:52:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:52:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:52:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:52:58 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:52:59 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:52:59 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://www.cnproxy.com/proxy5.html>: HTTP status code is not handled or not allowed 2018-11-29 19:52:59 [cralwer] INFO: Read 10 requests from haipproxy:spider:ajax_gfw 2018-11-29 19:52:59 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=1>: HTTP status code is not handled or not allowed 2018-11-29 19:52:59 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://free-proxy.cz/en/proxylist/main/25>: HTTP status code is not handled or not allowed 2018-11-29 19:53:02 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=2>: HTTP status code is not handled or not allowed 2018-11-29 19:53:03 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://free-proxy.cz/en/proxylist/main/26>: HTTP status code is not handled or not allowed 2018-11-29 19:53:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:53:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:53:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:53:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:53:03 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:53:03 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:53:05 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=3>: HTTP status code is not handled or not allowed 2018-11-29 19:53:06 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://free-proxy.cz/en/proxylist/main/27>: HTTP status code is not handled or not allowed 2018-11-29 19:53:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:53:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:53:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:53:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:53:08 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:53:09 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:53:09 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://free-proxy.cz/en/proxylist/main/28>: HTTP status code is not handled or not allowed 2018-11-29 19:53:09 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=4>: HTTP status code is not handled or not allowed 2018-11-29 19:53:12 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://free-proxy.cz/en/proxylist/main/29>: HTTP status code is not handled or not allowed 2018-11-29 19:53:13 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 https://proxy-list.org/english/index.php?p=5>: HTTP status code is not handled or not allowed 2018-11-29 19:53:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http 2018-11-29 19:53:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:http:temp 2018-11-29 19:53:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:https:temp 2018-11-29 19:53:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:weibo:temp 2018-11-29 19:53:13 [cralwer] INFO: Read 0 ip proxies from haipproxy:zhihu:temp 2018-11-29 19:53:13 [cralwer] INFO: Read 10 requests from haipproxy:spider:ajax_gfw 2018-11-29 19:53:14 [cralwer] INFO: Read 0 requests from haipproxy:spider:gfw 2018-11-29 19:53:15 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://free-proxy.cz/en/proxylist/main/15>: HTTP status code is not handled or not allowed 2018-11-29 19:53:18 [scrapy.spidermiddlewares.httperror] INFO: Ignoring response <502 http://free-proxy.cz/en/proxylist/main/16>: HTTP status code is not handled or not allowed 2018-11-29 19:53:18 [cralwer] INFO: Read 0 ip proxies from haipproxy:init:http

从这个log上面看好像什么也没爬到,但是数据库里面明明有了好多爬到的IP,不是很清楚怎么回事。

by the way,我的树莓派是arm芯片的,而那个splash好像只有x86和x64的,我如果想直接在树莓派上运行haipproxy的话要怎么弄呢?

lmst2 commented 5 years ago

试了一下Ubuntu,也是同样的情况,就好像validator没有正常工作一样

ghost commented 5 years ago

我也遇到类似的情况,程序debug也麻烦,很难排查问题。