scrapy练习1

1.建立项目:

#建立名为tuto的项目
scrapy startproject tuto

2.进入项目目录:

cd tuto

3.建立域名任务:

#minyan任务名;后面是任务对应的域名
scrapy genspider minyan quotes.toscrape.com

cmd全程记录:

E:ascrapy_test>scrapy startproject tuto
New Scrapy project 'tuto', using template directory 'c:pythonpython37libsite
-packagesscrapy	emplatesproject', created in:
    E:ascrapy_test	uto

You can start your first spider with:
    cd tuto
    scrapy genspider example example.com

E:ascrapy_test>cd tuto

E:ascrapy_test	uto>scrapy genspider minyan quotes.toscrape.com
Created spider 'minyan' using template 'basic' in module:
  tuto.spiders.minyan

E:ascrapy_test	uto>

4.pycharm打开项目,建立用于调试的文件:main.py

用到的函数解析:https://www.cnblogs.com/chenxi188/p/10876690.html

 main.py:

# -*- coding: utf-8 -*-
__author__='pasaulis'
import sys,os
from scrapy.cmdline import execute

#获取到当前目录,添加到环境变量,方便运行cmd不必再去找目录
#.append()添加到环境变量;.dirname()输出不含文件名的路径;.abspath(__file__)得到当前文件的绝对路径
sys.path.append(os.path.dirname(os.path.abspath(__file__)))

#调用命令运行爬虫scrapy crawl 爬虫名minyan.py中(name的值)
execute(["scrapy","crawl","minyan"])

 查看页面,确定需要解析哪些数据,在items.py里写出:

import scrapy


class TutoItem(scrapy.Item):
    # define the fields for your item here like:
    # name = scrapy.Field()
    text=scrapy.Field()
    author=scrapy.Field()
    tags=scrapy.Field()
    #pass

minyan.py:

# -*- coding: utf-8 -*-
import scrapy
from tuto.items import TutoItem #从tuto文件夹下的items.py导入其中的类TutoItem用于创建列表

class MinyanSpider(scrapy.Spider):
    name = 'minyan'
    allowed_domains = ['quotes.toscrape.com']
    start_urls = ['http://quotes.toscrape.com/']

    def parse(self, response):
        quotes=response.css('.quote')
        for quote in quotes:
            item=TutoItem()
            item['text']=quote.css('.text::text').extract_first()
            item['author']=quote.css('.author::text').extract_first()
            item['tags']=quote.css('.keywords::attr(content)').extract_first()
            yield item

        #下一页
        next=response.css('.next > a::attr(href)').extract_first()
        url=response.urljoin(next) #补全网址
        yield scrapy.Request(url=url,callback=self.parse) #请求下一页,并回调自身方法

settings.py把此条改为false

# Obey robots.txt rules
ROBOTSTXT_OBEY = False

运行代码:

在pycharm最下terminal里: scrapy crawl mingyan

或main.py里运行

原文地址:https://www.cnblogs.com/chenxi188/p/10876901.html