Skip to content

RequestList and stateKeyPrefix #2440

Answered by janbuchar
tsrdatatech asked this question in Q&A
Discussion options

You must be logged in to vote

If you want to terminate a crawler and resume it, just disabling CRAWLEE_PURGE_ON_START should be enough. RequestList is not necessary, the default RequestQueue will probably work better.

If you want to process some URLs every time you run the crawler, even if you already processed them before, then giving them a random unique key is a good solution.

Replies: 2 comments 1 reply

Comment options

You must be logged in to vote
0 replies
Comment options

You must be logged in to vote
1 reply
@tsrdatatech
Comment options

Answer selected by tsrdatatech
Sign up for free to join this conversation on GitHub. Already have an account? Sign in to comment
Category
Q&A
Labels
None yet
2 participants