本文实例讲述了python实现批量下载新浪博客的方法。分享给大家供大家参考。具体实现方法如下:

# coding=utf-8 
import urllib2
import sys, os
import re
import string
from BeautifulSoup import BeautifulSoup
def encode(s):
  return s.decode(\'utf-8\').encode(sys.stdout.encoding, \'ignore\')
def getHTML(url):
  #proxy_handler = urllib2.ProxyHandler({\'http\':\'http://211.138.124.211:80\'})
  #opener = urllib2.build_opener(proxy_handler)
  #urllib2.install_opener(opener)
  req = urllib2.Request(url)
  response = urllib2.urlopen(req, timeout=15)
  return BeautifulSoup(response, convertEntities=BeautifulSoup.HTML_ENTITIES)
def visible(element):
  \'\'\'抓取可见的文本元素\'\'\'
  if element.parent.name in [\'style\', \'script\', \'[document]\', \'head\', \'title\']:
    return False
  elif re.match(\'\', str(element)):
    return False
  elif element == u\'\\xa0\':
    return False
  return True
def delReturn(element):
  \'\'\'删除元素内的换行\'\'\'
  return re.sub(\'(?\"|\\xa0]\', \'\', filename)
def writeToFile(text, filename, dirname):
  if not os.path.exists(dirname):
    os.makedirs(dirname)
    print encode(\'保存到目录\'), dirname
  filename = validFilename(filename)
  print encode(\'保存文章\'), filename
  path = os.path.join(dirname, filename)
  if not os.path.exists(path):
    f = open(path, \'w\')
    f.write(text)
    f.close()
  else:
    print filename, encode(\'已经存在\')
def formatContent(url, title=\'\'):
  \'\'\'格式化文章内容\'\'\'
  page = getHTML(url)
  content = page.find(\'div\', {\'class\':\'articalContent\'})
  art_id = re.search(\'blog_(\\w+)\\.html\', url).group(1)
  blog_name = page.find(\'span\', id=\'blognamespan\').string
  if title == \'\':
    title = page.find(\'h2\', id=re.compile(\'^t_\')).string
  temp_data = filter(visible, content.findAll(text=True)) # 去掉不可见元素
  temp_data = \'\'.join(map(delReturn, temp_data)) # 删除元素内的换行符
  temp_data = temp_data.strip() # 删除文章首尾的空行
  temp_data = re.sub(\'\\n{2,}\', \'\\n\\n\', temp_data) # 删除文章内过多的空行
  # 输出到文件
  # 编码问题
  temp_data = \'本文地址:\'.decode(\'utf-8\') + url + \'\\n\\n\' + temp_data
  op_text = temp_data.encode(\'utf-8\')
  op_file = title + \'_\' + art_id +\'.txt\'
  writeToFile(op_text, op_file, blog_name)
def articlelist(url):
  articles = {}
  page = getHTML(url)
  pages = page.find(\'ul\', {\'class\':\'SG_pages\'}).span.string
  page_num = int(re.search(\'(\\d+)\', pages).group(1))
  for i in range(1, page_num+1):
    print encode(\'生成第%d页文章索引\'%i)
    if i != 1:
      url = re.sub(\'(_)\\d+(\\.html)$\', \'\\g<1>\'+str(i)+\'\\g<2>\', url)
      page = getHTML(url)
    article = page.findAll(\'span\', {\'class\':\'atc_title\'})
    for art in article:
      art_title = art.a[\'title\']
      art_href = art.a[\'href\']
      articles[art_title] = art_href
  return articles
def blog_dld(articles):
  if not isinstance(articles, dict):
    return False
  print encode(\'开始下载文章\')
  for art_title, art_href in articles.items():
    formatContent(art_href, art_title)
if __name__ == \'__main__\':
  sel = raw_input(encode(\'你要下载的是(1)全部文章还是(2)单篇文章,输入1或者2: \'))
  if sel == \'1\':
    #articlelist_url = \'http://blog.sina.com.cn/s/articlelist_1303481411_0_1.html\'
    articlelist_url = raw_input(encode(\'请输入博客文章目录链接: \'))
    articles = articlelist(articlelist_url)
    blog_dld(articles)
  else:
    #article_url = \'http://blog.sina.com.cn/s/blog_4db18c430100gxc5.html\'
    article_url = raw_input(encode(\'请输入博客文章链接: \'))
    formatContent(article_url)

希望本文所述对大家的Python程序设计有所帮助。