Closed josselinlbe closed 6 years ago
Without a persistent distributed cookie cache like as discussed in this comment, your crawls on your logged in page will not have the same cookie token being passed on different requests for a depth > 0.
If you are comfortable having each spider log in upon startup, this may work, however you should look at the cookie implementation in the project to make sure it will work like you expect.
I presume your use case is to have your entire cluster log in to a specific website, and then on-demand crawl pages within the site? It is an interesting use case but would limit this "generalized" project so I will have to think about it to see if it makes sense to have the capability here.
Otherwise, since this is a personal setup and not a bug or issue, can we close this and move to the gitter chat room?
Closing. I think this is best reserved for a personal/custom setup and not a generic setup as supported by this project.
Hello ! i want to create my own spider with the function to simulate an user login. But i don't understand this init:
How to had the login system with LinkSpider ? Here is my beginning code... Do you think I can fit it in?
Thanks :+1: