With python3.7 it produces same output Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:17 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:19 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:24 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes cannot import name 'etree' from 'lxml' (/usr/lib/python3/dist-packages/lxml/__init__.py)Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:29 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:34 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:39 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:44 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:49 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:54 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:25:59 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:26:04 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:26:09 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:26:14 [scrapy.extensions.logstats] INFO: Crawled 0 pages (at 0 pages/min), scraped 0 items (at 0 items/min) Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes 2019-01-04 14:26:14 [twisted] CRITICAL: Unhandled Error Traceback (most recent call last): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/commands/crawl.py", line 58, in run self.crawler_process.start() File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/crawler.py", line 293, in start reactor.run(installSignalHandlers=False) # blocking call File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1267, in run self.mainLoop() File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 1276, in mainLoop self.runUntilCurrent() --- --- File "/usr/local/lib/python3.7/dist-packages/twisted/internet/base.py", line 902, in runUntilCurrent call.func(*call.args, **call.kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/utils/reactor.py", line 41, in __call__ return self._func(*self._a, **self._kw) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 135, in _next_request self.crawl(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 210, in crawl self.schedule(request, spider) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/engine.py", line 216, in schedule if not self.slot.scheduler.enqueue_request(request): File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 57, in enqueue_request dqok = self._dqpush(request) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 86, in _dqpush self.dqs.push(reqd, -request.priority) File "/usr/lib/python3/dist-packages/queuelib/pqueue.py", line 33, in push self.queues[priority] = self.qfactory(priority) File "/usr/local/lib/python3.7/dist-packages/Scrapy-1.5.0-py3.7.egg/scrapy/core/scheduler.py", line 114, in _newdq return self.dqclass(join(self.dqdir, 'p%s' % priority)) File "/usr/lib/python3/dist-packages/queuelib/queue.py", line 142, in __init__ self.size, = struct.unpack(self.SIZE_FORMAT, qsize) struct.error: unpack requires a buffer of 4 bytes