Using python to make beautiful image Crawlers,
The delayed loading technology is used for the loading of petal images. The source code can only download more than 20 images. After modification, the source code can basically download all the images, but the speed is a little slow and will be optimized later.
import urllib, urllib2, re, sys, os,requestspath=r"C:\wqa\beautify"url = 'http://huaban.com/favorite/beauty'#http://huaban.com/explore/zhongwenlogo/?ig1un9tq&max=327773629&limit=20&wfl=1i_headers = {"User-Agent": "Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/46.0.2490.71 Safari/537.36"}count=0def urlHandle(url): req = urllib2.Request(url, headers=i_headers) html = urllib2.urlopen(req).read() reg = re.compile(r'"pin_id":(\d+),.+?"file":{"farm":"farm1", "bucket":"hbimg",.+?"key":"(.*?)",.+?"type":"image/(.*?)"', re.S) groups = re.findall(reg, html) return groupsdef imgHandle(groups): if groups: for att in groups: pin_id = att[0] att_url = att[1] + '_fw236' img_type = att[2] img_url = 'http://img.hb.aicdn.com/' + att_url r = requests.get(img_url) with open(path + att_url + '.' + img_type, 'wb') as fd: for chunk in r.iter_content(): fd.write(chunk)groups = urlHandle(url)imgHandle(groups)while(groups): count+=1 print count pin_id = groups[-1][0] print pin_id urltemp = url+'/?max=' + str(pin_id) + '&limit=' + str(20) + '&wfl=1' print(urltemp) groups = urlHandle(urltemp) #print groups imgHandle(groups)