Create workers 1
In this section we will write the code that will enable multiple spiders to run at the same time and crawl the page simultaneously. Write the following code at the end of the main.py file:
def create_jobs(): # this function is called as long there are links that need to be crawled
for link in file_to_set(QUEUE_FILE):
queue.put(link) # stores the link in the thread queue
queue.join()
crawl() # calls the crawl() function to get the update version
def crawl(): # this function will check if there are items in the queue, and if there are, it will crawl them
queued_links = file_to_set(QUEUE_FILE) # converts to set
if len(queued_links) > 0: # checks if there are items that need to be crawled
print(str(len(queued_links)) + ' links in the queue') # prints the info message
create_jobs()



