mirror of
https://github.com/scrapy/scrapy.git
synced 2025-02-21 06:33:07 +00:00
tests for new signal
This commit is contained in:
parent
ae04174884
commit
9916f6e556
59
tests/test_request_left.py
Normal file
59
tests/test_request_left.py
Normal file
@ -0,0 +1,59 @@
|
|||||||
|
from twisted.internet import defer
|
||||||
|
from twisted.trial.unittest import TestCase
|
||||||
|
from scrapy.signals import request_left_downloader
|
||||||
|
from scrapy.spiders import Spider
|
||||||
|
from scrapy.utils.test import get_crawler
|
||||||
|
from tests.mockserver import MockServer
|
||||||
|
|
||||||
|
class SignalCatcherSpider(Spider):
|
||||||
|
name = 'signal_catcher'
|
||||||
|
|
||||||
|
def __init__(self, crawler, url, *args, **kwargs):
|
||||||
|
super(SignalCatcherSpider, self).__init__(*args, **kwargs)
|
||||||
|
crawler.signals.connect(self.on_response_download,
|
||||||
|
signal=request_left_downloader)
|
||||||
|
self.catched_times = 0
|
||||||
|
self.start_urls = [url]
|
||||||
|
|
||||||
|
@classmethod
|
||||||
|
def from_crawler(cls, crawler, *args, **kwargs):
|
||||||
|
spider = cls(crawler, *args, **kwargs)
|
||||||
|
return spider
|
||||||
|
|
||||||
|
def on_response_download(self, request, spider):
|
||||||
|
self.catched_times = self.catched_times + 1
|
||||||
|
|
||||||
|
|
||||||
|
class TestCatching(TestCase):
|
||||||
|
|
||||||
|
def setUp(self):
|
||||||
|
self.mockserver = MockServer()
|
||||||
|
self.mockserver.__enter__()
|
||||||
|
|
||||||
|
def tearDown(self):
|
||||||
|
self.mockserver.__exit__(None, None, None)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def test_success(self):
|
||||||
|
crawler = get_crawler(SignalCatcherSpider)
|
||||||
|
yield crawler.crawl(self.mockserver.url("/status?n=200"))
|
||||||
|
self.assertEqual(crawler.spider.catched_times, 1)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def test_timeout(self):
|
||||||
|
crawler = get_crawler(SignalCatcherSpider,
|
||||||
|
{'DOWNLOAD_TIMEOUT': 0.1})
|
||||||
|
yield crawler.crawl(self.mockserver.url("/delay?n=0.2"))
|
||||||
|
self.assertEqual(crawler.spider.catched_times, 1)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def test_disconnect(self):
|
||||||
|
crawler = get_crawler(SignalCatcherSpider)
|
||||||
|
yield crawler.crawl(self.mockserver.url("/drop"))
|
||||||
|
self.assertEqual(crawler.spider.catched_times, 1)
|
||||||
|
|
||||||
|
@defer.inlineCallbacks
|
||||||
|
def test_noconnect(self):
|
||||||
|
crawler = get_crawler(SignalCatcherSpider)
|
||||||
|
yield crawler.crawl('http://thereisdefinetelynosuchdomain.com')
|
||||||
|
self.assertEqual(crawler.spider.catched_times, 1)
|
Loading…
x
Reference in New Issue
Block a user