Simple collection of crawlers
# 1. I only made a function to collect URLs. If you need to store the data into the database, you can write the processing code in the parseData function.
# 2. Requires sqlite3 or pysqlite support
# 3. Can run on DreamHost.com space
# 4. User-Agent can be modified to pretend to be a search engine spider
# 5. You can set the pause time and control the collection speed.
Expand