BeautifulSoup4小试牛刀

2016-08-03  本文已影响49人  Snow_DZG

python真好玩,python真有趣,python真该死!尝试爬取起点中文网的小说,因为他家的小说网页比较复杂,有一点挑战性!

问题(慢慢解决):
1.代码杂乱,可以适当精简重构
2.可以采用多线程的方式去爬,效率可以更高
3.适当的采用各种包各种方式,让网站不认为自己是爬虫

不多说直接上代码:

需要注意:
1."C:/Users/DZG/Desktop/" 这个路径换成自己本机的
2.爬取的url,可以复制一个去看下 这个url是哪里的url,-~

#! usr/bin/env python3
# -*- coding:utf-8 -*-
# author DZG 2016-08-03

#问题(慢慢解决):
#1.代码杂乱,可以适当精简重构
#2.可以采用多线程的方式去爬,效率可以更高
#3.适当的采用各种包各种方式,让网站不认为自己是爬虫
#4.目前只是爬取免费,后期应该可以加入 模拟VIP登录的功能
#需要注意:
#1."C:/Users/DZG/Desktop/" 这个路径换成自己本机的
#2.爬取的url,可以复制一个去看下 这个url是哪里的url,^-^~

import urllib.request
import chardet
import time
import random
from bs4 import BeautifulSoup as bs

ISOTIMEFORMAT='%Y-%m-%d %X'

#通用装饰器1
def writeBook(func):
    def swr(*args,**kw):
        f = open('C:/Users/DZG/Desktop/' + args[-1] + '.md', 'a', encoding='utf-8');
        f.write('\n\n')
        result = func(*args, **kw)
        f.write('>')
        return result;
    return swr;

#创建文件并定义名称
def createBook(bookName):
    f=open('C:/Users/DZG/Desktop/'+ bookName +'.md','a',encoding='utf-8');

#写卷目录
def createTitle(title,bookName):
    f=open('C:/Users/DZG/Desktop/'+bookName+'.md','a',encoding='utf-8');
    f.write('\n\n### ' + title)

#写每卷中对应的章节目录
def writeZhangjie(menuContentBs, bookName,i):
    menuZhangJie = menuContentBs.find_all("div", "box_cont")[1:];
    menuZhangJieS = bs(str(menuZhangJie[i]), 'html.parser');
    menuZhangJieSpan = menuZhangJieS.find_all('span')
    menuZhangJieHref = menuZhangJieS.find_all('a')
    menuZhangJieHrefList = getHref(menuZhangJieHref)
    writeZhangjieDetail(menuZhangJieSpan, bookName,menuZhangJieHrefList)

def writeZhangjie2(menuContentBs, bookName,i):
    menuZhangJie = menuContentBs.find_all("div", "box_cont")[0:];
    menuZhangJieS = bs(str(menuZhangJie[i]), 'html.parser');
    menuZhangJieSpan = menuZhangJieS.find_all('span')
    menuZhangJieHref = menuZhangJieS.find_all('a')
    menuZhangJieHrefList = getHref(menuZhangJieHref)
    writeZhangjieDetail(menuZhangJieSpan, bookName,menuZhangJieHrefList)

#写章节细节和每章内容的方法(还有点问题)
def writeZhangjieDetail(menuZhangJieSpan, bookName,menuZhangJieHrefList):
    j = 0
    for spanValue in menuZhangJieSpan:
        timeout = random.choice(range(80, 180)) / 100
        print('延时:', timeout)

        spanValue = bs(str(spanValue), 'html.parser').get_text()

        print(j, spanValue, '--执行到此章节--', time.strftime(ISOTIMEFORMAT, time.localtime())) #打印已经写到哪个章节
        f = open('C:/Users/DZG/Desktop/' + bookName + '.md', 'a', encoding='utf-8');
        f.write('\n\n---')
        f.write('\n\n#### ' + spanValue + '\n\n')
        try:
            chapterCode = urllib.request.urlopen(menuZhangJieHrefList[j]).read()
            result = chardet.detect(chapterCode)  # 检验读取的页面的编码方式
            if (result['confidence'] > 0.5):  # 如果概率大于0.5 即采取这种编码
                chapterCode = chapterCode.decode(result['encoding'])
            chapterSoup = bs(chapterCode, 'html.parser')  # 使用BS读取解析网页代码
            chapterResult = chapterSoup.find(id='chaptercontent')  # 找到id='chaptercontent'的节点

            chapterResultChilds = chapterResult.children
            storySrc = bs(str(list(chapterResultChilds)[3]), "html.parser")
            fileurl = storySrc.contents[0].attrs['src']

            fileArticle1 = urllib.request.urlopen(fileurl).read().decode('gb2312', errors='replace')
            fileArticle2 = bs(fileArticle1, 'html.parser');
            fileArticle = str(fileArticle2).replace("document.write('  ", "")\
                .replace("<p>  ", "  \n\n").replace("</p>", "").replace("');","")\
                .replace('<a href="http://www.qidian.com">起点'
                         '中文网 www.qidian.com 欢迎广大书友光临阅读,最新、最快、最火的连载作品'
                         '尽在起点原创!</a><a>手机用户请到m.qidian.com阅读。</a>',"---来自DZG")\
                .replace('<ahref=http: www.qidian.com="">起点'
                         '中文网www.qidian.com欢迎广大书友光临阅读,最新、最快、最火的连载作品'
                         '尽在起点原创!</ahref=http:><a>手机用户请到m.qidian.com阅读。</a>',"---来自DZG")\
                .replace("起点中文网www.qidian.com欢迎广大书友光临阅读,最新、最快、最火的连载作品"
                         "尽在起点原创!</a><a>手机用户请到m.qidian.com阅读。</a></a>","---来自DZG")\
                .replace("http://www.qidian.com","http://www.jianshu.com/users/6ca20b30c14c");
            f.write(fileArticle)
        except:
            f.write('章节丢失')
            print('章节丢失')
        j = j + 1
        time.sleep(timeout)

#获取各个章节的URL
def getHref(menuZhangJieHref):
    menuZhangJieHrefList = []
    for index in menuZhangJieHref:
        menuZhangJieHrefList.append(index.get('href'));
    return menuZhangJieHrefList;

def write(menuContentBs,menu,n):
    i = 0
    for index in menu:
        MuLu = bs(str(index), 'html.parser');
        MLb = MuLu.find('b');
        MuLua = bs(str(MLb), 'html.parser');
        MLa = MuLua.find('a');
        forTitle = bs(str(MLa), 'html.parser');
        muluTitle = MuLua.get_text()
        bookName = forTitle.get_text()
        createBook(bookName)
        createTitle(muluTitle, bookName)
        # MD文件和卷标题写好,就要写入正式的章节目录和文本内容
        if n==0:
            writeZhangjie2(menuContentBs, bookName, i)
        else:
            writeZhangjie(menuContentBs, bookName, i)
        print(i + 1, muluTitle, '--执行到此目录', time.strftime(ISOTIMEFORMAT, time.localtime()))  # 打印已经写入的目录
        i = i + 1;
        # time.sleep(1)

def main():
    #url = "http://read.qidian.com/BookReader/uMyvtX3MUFw1.aspx"  # 日在火影
    #url = "http://read.qidian.com/BookReader/7FnsrjmItYk1.aspx"  # 大时代1958
    url = "http://read.qidian.com/BookReader/nxkadQuAGmYJiWg6PYdjVg2.aspx"  # 绿茵圣骑士

    webStory = urllib.request.urlopen(url).read().decode('utf8',errors='replace');          #获取整个网页

    print('开始:',time.strftime( ISOTIMEFORMAT, time.localtime() ))
    soup = bs(webStory,'html.parser')                       #解析整个网页
    menuContent = soup.find_all(id="content");              #获取解析好的网页上id位content的元素
    #menuContentStr = ','.join(str(v) for v in menuContent)  #将 menuContent转换为str 如果有多个
    menuContentBs = bs(str(menuContent),'html.parser');       #解析转换好的menuContentStr为bs对象

    menu = menuContentBs.find_all("div","box_title")[0:1]
    if('正文卷' in bs(str(menu),'html.parser').get_text()):
        menu = menuContentBs.find_all("div", "box_title")[0:];  # 从解析好的bs对象中获取是div且class为box_title的元素
        write(menuContentBs,menu, 0)
    else:
        menu = menuContentBs.find_all("div", "box_title")[1:];  # 从解析好的bs对象中获取是div且class为box_title的元素
        write(menuContentBs,menu, 1)
    #soup2 = bs(str(menu),'html.parser');                       #转换为bs对象

    print('结束:',time.strftime( ISOTIMEFORMAT, time.localtime() ))

#主函数
if __name__ == '__main__':
    main();

下面是爬的样品:
http://www.jianshu.com/p/a7ae79bd721a

上一篇 下一篇

猜你喜欢

热点阅读