scrapy中的debug信息
在scrapy中设置log
1、在settings中设置log级别,在settings.py中添加一行:
Scrapy提供5层logging级别: CRITICAL - 严重错误(critical) ERROR - 一般错误(regular errors) WARNING - 警告信息(warning messages) INFO - 一般信息(informational messages) DEBUG - 调试信息(debugging messages)
scrapy默认显示DEBUG级别的log信息
2、将输出的结果保存为log日志,在settings.py中添加路径:
LOG_FILE = ‘./log.log‘
3、显示log位置,在pipelines.py中:
import logging logger = logging.getLogger(__name__) def process_item(self, item, spider): logger.warning(item) ....
4.在spider
文件中引入Log日志:
class DcdappSpider(scrapy.Spider): name = ‘dcdapp‘ allowed_domains = [‘m.dcdapp.com‘] custom_settings = { # 设置管道下载 ‘ITEM_PIPELINES‘: { ‘autospider.pipelines.DcdAppPipeline‘: 300, }, # 设置log日志 ‘LOG_LEVEL‘:‘DEBUG‘, ‘LOG_FILE‘:‘./././Log/dcdapp_log.log‘ }
原文地址:https://www.cnblogs.com/youxiu123/p/11624506.html
时间: 2024-11-23 16:29:55