From 6392531b0901f5a1546424fdac432b71fb157a47 Mon Sep 17 00:00:00 2001 From: guange <8863824@gmail.com> Date: Thu, 17 Jan 2019 22:48:39 +0800 Subject: [PATCH] . --- chapter1/crawler/taobao/settings.py | 123 ---------------------------- 1 file changed, 123 deletions(-) delete mode 100644 chapter1/crawler/taobao/settings.py diff --git a/chapter1/crawler/taobao/settings.py b/chapter1/crawler/taobao/settings.py deleted file mode 100644 index d73d1b6..0000000 --- a/chapter1/crawler/taobao/settings.py +++ /dev/null @@ -1,123 +0,0 @@ -# -*- coding: utf-8 -*- - -# Scrapy settings for taobao project -# -# For simplicity, this file contains only settings considered important or -# commonly used. You can find more settings consulting the documentation: -# -# https://doc.scrapy.org/en/latest/topics/settings.html -# https://doc.scrapy.org/en/latest/topics/downloader-middleware.html -# https://doc.scrapy.org/en/latest/topics/spider-middleware.html - -BOT_NAME = 'taobao' - -SPIDER_MODULES = ['taobao.spiders'] -NEWSPIDER_MODULE = 'taobao.spiders' - - - - -# Crawl responsibly by identifying yourself (and your website) on the user-agent -USER_AGENT = 'Mozilla/5.0 (Macintosh; Intel Mac OS X 10_14_1) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/70.0.3538.102 Safari/537.36' - -# Obey robots.txt rules -ROBOTSTXT_OBEY = False - -LOG_LEVEL='DEBUG' - - -SPLASH_URL = 'http://127.0.0.1:8050' - -# Configure maximum concurrent requests performed by Scrapy (default: 16) -#CONCURRENT_REQUESTS = 32 - -# Configure a delay for requests for the same website (default: 0) -# See https://doc.scrapy.org/en/latest/topics/settings.html#download-delay -# See also autothrottle settings and docs -#DOWNLOAD_DELAY = 3 -# The download delay setting will honor only one of: -#CONCURRENT_REQUESTS_PER_DOMAIN = 16 -#CONCURRENT_REQUESTS_PER_IP = 16 - -# Disable cookies (enabled by default) -COOKIES_ENABLED = True - -# Disable Telnet Console (enabled by default) -TELNETCONSOLE_ENABLED = True - -# Override the default request headers: -#DEFAULT_REQUEST_HEADERS = { -# 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', -# 'Accept-Language': 'en', -#} - - -# Enable or disable spider middlewares -# See https://doc.scrapy.org/en/latest/topics/spider-middleware.html -#SPIDER_MIDDLEWARES = { -# 'taobao.middlewares.ZhihuSpiderMiddleware': 543, -#} - -# Enable or disable downloader middlewares -# See https://doc.scrapy.org/en/latest/topics/downloader-middleware.html -#DOWNLOADER_MIDDLEWARES = { -# 'taobao.middlewares.ZhihuDownloaderMiddleware': 543, -#} - -DOWNLOADER_MIDDLEWARES = { - 'scrapy_splash.SplashCookiesMiddleware': 723, - 'scrapy_splash.SplashMiddleware': 725, - 'scrapy.downloadermiddlewares.httpcompression.HttpCompressionMiddleware': 810, -} -#Splash自己的去重过滤器 -DUPEFILTER_CLASS = 'scrapy_splash.SplashAwareDupeFilter' - - - -# Enable or disable extensions -# See https://doc.scrapy.org/en/latest/topics/extensions.html -#EXTENSIONS = { -# 'scrapy.extensions.telnet.TelnetConsole': None, -#} - -# Configure item pipelines -# See https://doc.scrapy.org/en/latest/topics/item-pipeline.html -#ITEM_PIPELINES = { -# 'taobao.pipelines.ZhihuPipeline': 300, -#} - -ITEM_PIPELINES = {'taobao.pipelines.MyImagesPipeline': 300} - -IMAGES_STORE = '/tmp/taobao' - -# Enable and configure the AutoThrottle extension (disabled by default) -# See https://doc.scrapy.org/en/latest/topics/autothrottle.html -# 启用限速 -AUTOTHROTTLE_ENABLED = True -# The initial download delay -AUTOTHROTTLE_START_DELAY = 30 -# The maximum download delay to be set in case of high latencies -AUTOTHROTTLE_MAX_DELAY = 60 -# The average number of requests Scrapy should be sending in parallel to -# each remote server -AUTOTHROTTLE_TARGET_CONCURRENCY = 1.0 -# Enable showing throttling stats for every response received: -#AUTOTHROTTLE_DEBUG = False - -# Enable and configure HTTP caching (disabled by default) -# See https://doc.scrapy.org/en/latest/topics/downloader-middleware.html#httpcache-middleware-settings -#HTTPCACHE_ENABLED = True -#HTTPCACHE_EXPIRATION_SECS = 0 -#HTTPCACHE_DIR = 'httpcache' -#HTTPCACHE_IGNORE_HTTP_CODES = [] -#HTTPCACHE_STORAGE = 'scrapy.extensions.httpcache.FilesystemCacheStorage' - -FEED_EXPORT_ENCODING='utf-8' - - -#HBASE_HOST = '106.75.85.84' -HBASE_HOST = '106.75.120.249' -HBASE_PORT = 40009 -HBASE_TABLE = 'jd' - -COMMENT_MAX_PAGE = 10