Size: a a a

2020 March 10

AK

Anil Kumar in Scrapy
I have 1000 pages to follow and each page contains 40 links to the pages which I have to scrape data from. Which spider you guys suggest? Single domain only.
источник

AR

Andrey Rahmatullin in Scrapy
Spider.
источник

AR

Andrey Rahmatullin in Scrapy
the page counts don't matter in any case
источник

i

ildar in Scrapy
Anil Kumar
How to set retry times for a particular request in crawl spider?
just for curiosity, why do you need this?
источник

AK

Anil Kumar in Scrapy
ildar
just for curiosity, why do you need this?
I'm doing freelancing job. I need it
источник

AK

Anil Kumar in Scrapy
I didn't get
источник

AR

Andrey Rahmatullin in Scrapy
Anil Kumar
I didn't get
what answer did you want then?
источник

i

ildar in Scrapy
ehm, I'm asking why you have to set it for every request "manually"?
источник

AK

Anil Kumar in Scrapy
Scrapy normal spider or crawl spider?
источник

AR

Andrey Rahmatullin in Scrapy
Spider is the "normal spider"
источник

i

ildar in Scrapy
selecting between depends only on this thing - do you need to use Rule objects? If yes, use CrawlSpider.
источник

AK

Anil Kumar in Scrapy
ildar
ehm, I'm asking why you have to set it for every request "manually"?
For pagination If I get error. I'll not get the remaining data. Thats why
источник

AR

Andrey Rahmatullin in Scrapy
set it globally.
источник

i

ildar in Scrapy
you know, in most cases, scrapy will do this for you, automatically!
источник

i

ildar in Scrapy
you don't need to bother about this until encounter it, I suppose even default settings will do the job.
источник

AK

Anil Kumar in Scrapy
ildar
you don't need to bother about this until encounter it, I suppose even default settings will do the job.
I got the error. That's why I'm asking
источник

i

ildar in Scrapy
in particular request and every other go without problems?
источник

AK

Anil Kumar in Scrapy
I'm taking about only paginated links
источник

i

ildar in Scrapy
Then set max_retry_times on this particular request, see here: https://doc.scrapy.org/en/latest/topics/downloader-middleware.html#retry-times
источник

МС

Михаил Синегубов in Scrapy
народ, может у него ajax какой с кастомным заголовком?
пускай сайт покажет... если это не военная тайна
источник