Skip to content

Commit

Permalink
Updated and cleaned up imports to comply with Scrapy 1.0.
Browse files Browse the repository at this point in the history
  • Loading branch information
rmax committed Jul 6, 2015
1 parent f5d2d29 commit f3468d4
Show file tree
Hide file tree
Showing 9 changed files with 30 additions and 26 deletions.
5 changes: 3 additions & 2 deletions example-project/example/items.py
Original file line number Diff line number Diff line change
Expand Up @@ -4,8 +4,9 @@
# http://doc.scrapy.org/topics/items.html

from scrapy.item import Item, Field
from scrapy.contrib.loader import ItemLoader
from scrapy.contrib.loader.processor import MapCompose, TakeFirst, Join
from scrapy.loader import ItemLoader
from scrapy.loader.processors import MapCompose, TakeFirst, Join


class ExampleItem(Item):
name = Field()
Expand Down
14 changes: 9 additions & 5 deletions example-project/example/spiders/dmoz.py
Original file line number Diff line number Diff line change
@@ -1,17 +1,21 @@
from scrapy.linkextractors import LinkExtractor
from scrapy.selector import Selector
from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor
from scrapy.contrib.spiders import CrawlSpider, Rule
from scrapy.spiders import CrawlSpider, Rule

from example.items import ExampleLoader


class DmozSpider(CrawlSpider):
name = 'dmoz'
allowed_domains = ['dmoz.org']
start_urls = ['http://www.dmoz.org/']

categories_lx = LinkExtractor(restrict_xpaths='//div[@id="catalogs"]')
directory_lx = LinkExtractor(restrict_xpaths='//ul[@class="directory dir-col"]')

rules = (
Rule(SgmlLinkExtractor(restrict_xpaths='//div[@id="catalogs"]')),
Rule(SgmlLinkExtractor(restrict_xpaths='//ul[@class="directory dir-col"]'),
callback='parse_directory', follow=True)
Rule(categories_lx),
Rule(directory_lx, callback='parse_directory', follow=True)
)

def parse_directory(self, response):
Expand Down
8 changes: 4 additions & 4 deletions example-project/example/spiders/mycrawler_redis.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
from scrapy_redis.spiders import RedisMixin
from scrapy.spiders import CrawlSpider, Rule
from scrapy.linkextractors import LinkExtractor

from scrapy.contrib.spiders import CrawlSpider, Rule
from scrapy.contrib.linkextractors.sgml import SgmlLinkExtractor
from scrapy_redis.spiders import RedisMixin

from example.items import ExampleLoader

Expand All @@ -13,7 +13,7 @@ class MyCrawler(RedisMixin, CrawlSpider):

rules = (
# follow all links
Rule(SgmlLinkExtractor(), callback='parse_page', follow=True),
Rule(LinkExtractor(), callback='parse_page', follow=True),
)

def set_crawler(self, crawler):
Expand Down
1 change: 1 addition & 0 deletions example-project/example/spiders/myspider_redis.py
Original file line number Diff line number Diff line change
@@ -1,4 +1,5 @@
from scrapy_redis.spiders import RedisSpider

from example.items import ExampleLoader


Expand Down
5 changes: 3 additions & 2 deletions scrapy_redis/dupefilter.py
Original file line number Diff line number Diff line change
@@ -1,9 +1,10 @@
import time
import connection

from scrapy.dupefilter import BaseDupeFilter
from scrapy.dupefilters import BaseDupeFilter
from scrapy.utils.request import request_fingerprint

from . import connection


class RFPDupeFilter(BaseDupeFilter):
"""Redis-based request duplication filter"""
Expand Down
7 changes: 3 additions & 4 deletions scrapy_redis/pipelines.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,7 @@
import redis
import connection

from twisted.internet.threads import deferToThread
from scrapy.utils.serialize import ScrapyJSONEncoder
from twisted.internet.threads import deferToThread

from . import connection


class RedisPipeline(object):
Expand Down
6 changes: 3 additions & 3 deletions scrapy_redis/scheduler.py
Original file line number Diff line number Diff line change
@@ -1,7 +1,7 @@
import connection

from scrapy.utils.misc import load_object
from scrapy_redis.dupefilter import RFPDupeFilter

from . import connection
from .dupefilter import RFPDupeFilter


# default values
Expand Down
7 changes: 3 additions & 4 deletions scrapy_redis/spiders.py
Original file line number Diff line number Diff line change
@@ -1,8 +1,7 @@
import connection

from scrapy import signals
from scrapy import Spider, signals
from scrapy.exceptions import DontCloseSpider
from scrapy.spider import Spider

from . import connection


class RedisMixin(object):
Expand Down
3 changes: 1 addition & 2 deletions scrapy_redis/tests.py
Original file line number Diff line number Diff line change
Expand Up @@ -3,8 +3,7 @@
import mock
import redis

from scrapy.http import Request
from scrapy.spider import Spider
from scrapy import Request, Spider
from unittest import TestCase

from . import connection
Expand Down

0 comments on commit f3468d4

Please sign in to comment.