From HERE I found that Scrapy engine fetch requests from scheduler before the start_urls
generated ones.
In my usage, I enqueued thousands of start urls (which from various domains) to the queue and the crawling goes not so fast (maybe networking issues). The problems comes up with me was that the spider itself extracts links and follows them, then Scrapy will fetch the requests from scheduler. It makes the concurrency lower.
I would like to learn about the design purpose of this mechanism.
BRs.