scrapyでプロジェクトを作成した後、crawlを実行しようとしたのですが、下記のようなエラーが出ました。青空文庫のあるhtmlのタイトルタグを抽出するプログラムなのですが、解決方法がわかる方、教えていただけると嬉しいです。
python
1import scrapy 2 3 4class SosekiSpider(scrapy.Spider): 5 name = 'soseki' 6 start_urls = [ 7 'https://www.aozora.gr.jp/index_pages/person148.html' 8 ] 9 10 def parse(self, response): 11 title = response.css('title') 12 print(title.extract()) 13
エラー
(base) C:\Users\rin20\PycharmProjects\pythonProject1\soseki_list\soseki_list\spiders>scrapy crawl soseki --nolog
Unhandled error in Deferred:
Traceback (most recent call last):
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\crawler.py", line 192, in crawl
return self._crawl(crawler, *args, **kwargs)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\crawler.py", line 196, in _crawl
d = crawler.crawl(*args, **kwargs)
File "C:\Users\rin20\anaconda3\lib\site-packages\twisted\internet\defer.py", line 1613, in unwindGenerator
return _cancellableInlineCallbacks(gen)
File "C:\Users\rin20\anaconda3\lib\site-packages\twisted\internet\defer.py", line 1529, in _cancellableInlineCallbacks
_inlineCallbacks(None, g, status)
--- <exception caught here> ---
File "C:\Users\rin20\anaconda3\lib\site-packages\twisted\internet\defer.py", line 1418, in _inlineCallbacks
result = g.send(result)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\crawler.py", line 87, in crawl
self.engine = self._create_engine()
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\crawler.py", line 101, in _create_engine
return ExecutionEngine(self, lambda : self.stop())
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\core\engine.py", line 69, in init
self.downloader = downloader_cls(crawler)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\core\downloader_init.py", line 83, in init
self.middleware = DownloaderMiddlewareManager.from_crawler(crawler)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\middleware.py", line 53, in from_crawler
return cls.from_settings(crawler.settings, crawler)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\middleware.py", line 35, in from_settings
mw = create_instance(mwcls, settings, crawler)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\utils\misc.py", line 167, in create_instance
instance = objcls.from_crawler(crawler, *args, **kwargs)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\downloadermiddlewares\robotstxt.py", line 36, in from_crawler
return cls(crawler)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\downloadermiddlewares\robotstxt.py", line 32, in init
self._parserimpl.from_crawler(self.crawler, b'')
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\robotstxt.py", line 124, in from_crawler
o = cls(robotstxt_body, spider)
File "C:\Users\rin20\anaconda3\lib\site-packages\scrapy\robotstxt.py", line 116, in init
from protego import Protego
builtins.ModuleNotFoundError: No module named 'protego'
回答2件
あなたの回答
tips
プレビュー
バッドをするには、ログインかつ
こちらの条件を満たす必要があります。
2020/12/17 15:16