Package: python-scrapy
Version: 0.22.2-1
Severity: grave
X-Debbugs-CC: Vincent Cheng <vch...@debian.org>

The tutorial example given at
http://doc.scrapy.org/en/0.22/intro/tutorial.html#intro-tutorial
doesn't work. If you do everything up to the "Crawling" section, this
happens:

# scrapy crawl dmoz
/home/kumar/Kumar/Programming/Python/Scrapy/tutorial/tutorial/spiders/dmoz_spider.py:3:
 ScrapyDeprecationWarning: tutorial.spiders.dmoz_spider.DmozSpider inherits 
from deprecated class scrapy.spider.BaseSpider, please inherit from 
scrapy.spider.Spider. (warning only on first subclass, there may be others)
  class DmozSpider(BaseSpider):
2014-03-31 08:08:44-0400 [scrapy] INFO: Scrapy 0.22.2 started (bot: tutorial)
2014-03-31 08:08:44-0400 [scrapy] INFO: Optional features available: ssl, 
http11, boto, django
2014-03-31 08:08:44-0400 [scrapy] INFO: Overridden settings: 
{'NEWSPIDER_MODULE': 'tutorial.spiders', 'SPIDER_MODULES': 
['tutorial.spiders'], 'BOT_NAME': 'tutorial'}
2014-03-31 08:08:44-0400 [scrapy] INFO: Enabled extensions: LogStats, 
TelnetConsole, CloseSpider, WebService, CoreStats, SpiderState
Traceback (most recent call last):
  File "/usr/bin/scrapy", line 4, in <module>
    execute()
  File "/usr/lib/python2.7/dist-packages/scrapy/cmdline.py", line 143, in 
execute
    _run_print_help(parser, _run_command, cmd, args, opts)
  File "/usr/lib/python2.7/dist-packages/scrapy/cmdline.py", line 89, in 
_run_print_help
    func(*a, **kw)
  File "/usr/lib/python2.7/dist-packages/scrapy/cmdline.py", line 150, in 
_run_command
    cmd.run(args, opts)
  File "/usr/lib/python2.7/dist-packages/scrapy/commands/crawl.py", line 50, in 
run
    self.crawler_process.start()
  File "/usr/lib/python2.7/dist-packages/scrapy/crawler.py", line 92, in start
    if self.start_crawling():
  File "/usr/lib/python2.7/dist-packages/scrapy/crawler.py", line 124, in 
start_crawling
    return self._start_crawler() is not None
  File "/usr/lib/python2.7/dist-packages/scrapy/crawler.py", line 139, in 
_start_crawler
    crawler.configure()
  File "/usr/lib/python2.7/dist-packages/scrapy/crawler.py", line 47, in 
configure
    self.engine = ExecutionEngine(self, self._spider_closed)
  File "/usr/lib/python2.7/dist-packages/scrapy/core/engine.py", line 62, in 
__init__
    self.scheduler_cls = load_object(self.settings['SCHEDULER'])
  File "/usr/lib/python2.7/dist-packages/scrapy/utils/misc.py", line 42, in 
load_object
    raise ImportError("Error loading object '%s': %s" % (path, e))
ImportError: Error loading object 'scrapy.core.scheduler.Scheduler': No module 
named queuelib

Please let me know if I need to provide additional details.

Thanks.

Kumar
-- 
Kumar Appaiah


-- 
To UNSUBSCRIBE, email to debian-bugs-dist-requ...@lists.debian.org
with a subject of "unsubscribe". Trouble? Contact listmas...@lists.debian.org

Reply via email to