This post was last edited by zzf.pdf at 22:36:33, January 24 ,.
Recently, I need to capture the content of a website. I used snoopy to capture the content. When I first found that an ip address was blocked, I switched the user-agent to google's spider according to the solution on the Internet, and forged an ip address with snoopy (a random ip address is changed for every capture) however, after capturing more than one hundred pages, it is still blocked by the ip address, and thus cannot be crawled. Is there any good solution?
Reply to discussion (solution)
The access is too frequent.
The access is too frequent. What should we do? sleep, but the amount of data to be captured is large. if sleep is used, there is not much time to capture
Snoopy is used to forge an ip address.
This is a fool of yours ..
Can I forge an ip address? This...
Take care of who you are.
Snoopy is used to forge an ip address.
This is a fool of yours.