Size: a a a

2020 March 13

AR

Andrey Rahmatullin in Scrapy
и у амазона всегда несколько вёрсток, отдающихся рандомно, может из-за этого
источник

AR

Andrey Rahmatullin in Scrapy
да
источник

AR

Andrey Rahmatullin in Scrapy
можешь в пауке сохранить респонс в файл и открыть в браузере
источник

🎱

🎱 in Scrapy
спасибо, попробую и так и так
источник

А

Александр in Scrapy
а нельзя разве в параметрах url передать
источник

А

Александр in Scrapy
источник

AK

Anil Kumar in Scrapy
Andrey Rahmatullin
extract URLs from your results (if that's impossible then from logs) and compare two lists
How can I do that in the spider and scrape the missing urls?
источник

AR

Andrey Rahmatullin in Scrapy
you shouldn't do that in the spider, you should process the results outside it
источник

AR

Andrey Rahmatullin in Scrapy
Александр
а нельзя разве в параметрах url передать
каких параметрах
источник

AK

Anil Kumar in Scrapy
Is there a way to do it in the spider?
источник

AR

Andrey Rahmatullin in Scrapy
yes, just write the code that loads the old results or logs, processes them and compares that with the old URL list
источник

AR

Andrey Rahmatullin in Scrapy
which is suboptimal, of course
источник

AR

Andrey Rahmatullin in Scrapy
to do that correctly you should have initially written your spider with retrying in mind, but it's too late now
источник

AK

Anil Kumar in Scrapy
I have given retry times as 70 and download delay. Concurrent requests = 50
источник

AR

Andrey Rahmatullin in Scrapy
whatever
источник

AK

Anil Kumar in Scrapy
Anil Kumar
I have given retry times as 70 and download delay. Concurrent requests = 50
Download delay =0
источник

AR

Andrey Rahmatullin in Scrapy
it doesn't matter now how you ran your previous job, if you don't have a list of URLs to retry you need to make it by comparing all URLs with successful URLs, or by finding failed URLs in the logs, there is no other way
источник

AR

Andrey Rahmatullin in Scrapy
there would be a lot of options if you asked before running the initial job
источник

AK

Anil Kumar in Scrapy
Andrey Rahmatullin
there would be a lot of options if you asked before running the initial job
Daily I have to scrape 200000 urls. Now tell me the options which you got
источник

AK

Anil Kumar in Scrapy
I'll modify the scraper
источник