From d11c8595e67680baad002bedf08dfd91f347cfb2 Mon Sep 17 00:00:00 2001 From: Mikhail Korobov Date: Mon, 22 Sep 2014 04:29:22 +0600 Subject: [PATCH] drop support for CONCURRENT_REQUESTS_PER_SPIDER --- scrapy/core/downloader/__init__.py | 8 -------- 1 file changed, 8 deletions(-) diff --git a/scrapy/core/downloader/__init__.py b/scrapy/core/downloader/__init__.py index a5e62a75d..ecbf33039 100644 --- a/scrapy/core/downloader/__init__.py +++ b/scrapy/core/downloader/__init__.py @@ -49,14 +49,6 @@ def _get_concurrency_delay(concurrency, spider, settings): if hasattr(spider, 'download_delay'): delay = spider.download_delay - # TODO: remove for Scrapy 0.15 - c = settings.getint('CONCURRENT_REQUESTS_PER_SPIDER') - if c: - warnings.warn("CONCURRENT_REQUESTS_PER_SPIDER setting is deprecated, " - "use CONCURRENT_REQUESTS_PER_DOMAIN instead", ScrapyDeprecationWarning) - concurrency = c - # ---------------------------- - if hasattr(spider, 'max_concurrent_requests'): concurrency = spider.max_concurrent_requests