本文实例讲述了python妹子图简单爬虫实现方法。分享给大家供大家参考。具体如下:
#!/usr/bin/env python #coding: utf-8 import urllib import urllib2 import os import re import sys #显示下载进度 def schedule(a,b,c): \'\'\'\'\' a:已经下载的数据块 b:数据块的大小 c:远程文件的大小 \'\'\' per = 100.0 * a * b / c if per > 100 : per = 100 print \'%.2f%%\' % per #获取html源码 def getHtml(url): page = urllib.urlopen(url) html = page.read() return html #下载图片 def downloadImg(html, num, foldername): picpath = \'%s\' % (foldername) #下载到的本地目录 if not os.path.exists(picpath): #路径不存在时创建一个 os.makedirs(picpath) target = picpath+\'/%s.jpg\' % num myItems = re.findall(\'\',html,re.S) print \'Downloading image to location: \' + target urllib.urlretrieve(myItems[0], target, schedule) #正则匹配分页 def findPage(html): myItems = re.findall(\'(\\d*)\', html, re.S) return myItems.pop() #正则匹配列表 def findList(html): myItems = re.findall(\'.*?
\', html, re.S) return myItems #总下载 def totalDownload(modelUrl): listHtml5 = getHtml(modelUrl) listContent = findList(listHtml) for list in listContent: html = getHtml(\'http://www.mzitu.com/\' + str(list[0])) totalNum = findPage(html) for num in range(1, int(totalNum)+1): if num == 1: url = \'http://www.mzitu.com/\' + str(list[0]) html5 = getHtml(url) downloadImg(html5, str(num), str(list[1])) else: url = \'http://www.mzitu.com/\' + str(list[0]) + \'/\'+str(num) html5 = getHtml(url) downloadImg(html5, str(num), str(list[1])) if __name__ == \'__main__\': listHtml = getHtml(\'http://www.mzitu.com/model\') #这是其中一个模块的url,可以添加不同的模块url从而达到整站爬取。 for model in range(1, int(findPage(listHtml))+1): if model == 1: modelUrl = \'http://www.mzitu.com/model\' totalDownload(modelUrl) else: modelUrl = \'http://www.mzitu.com/model/page/\' + str(model) totalDownload(modelUrl) print \"Download has finished.\"
希望本文所述对大家的Python程序设计有所帮助。
上一篇:python操作redis的方法