The so-called crawler is to get the source code of the webpage first, and then filter out the resources you want from the source code, such as pictures, videos and other files on the webpage, and even the text on the webpage. Next, we use Python to capture the pictures on the web page.
First, we get the source code of the website.
Then I analyze the resources I want from thousands of source codes. What I want here is the picture on the website.
Personally, this crawler tests the basic skills of regular expressions. How to write a regular expression, you can parse out all the resources you want, and the rest are relatively simple.
Here are some pictures I climbed down from the web page.