zoukankan      html  css  js  c++  java
  • 线程池 爬取一本小说

    #  本例子 开100线程

    import requests
    from lxml import etree
    from multiprocessing.dummy import Pool  # 线程类
    import time
    
    start  = time.time()
    
    url = "https://www.biquge.info/41_41486/"
    headers = {
        "user-agent": "Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.86 Safari/537.36"
    }
    page = requests.get(url=url, headers=headers)
    page.encoding = 'utf-8'
    page_text = page.text
    tree = etree.HTML(page_text)
    book_name = tree.xpath("//div[@id='info']/h1/text()")[0]  #小说书名
    author = tree.xpath("//div[@id='info']/p[1]/text()")[0].replace(' ', '').strip()
    book_type = tree.xpath("//div[@id='info']/p[2]/text()")[0].replace(' ', '')
    last_update_time = tree.xpath("//div[@id='info']/p[3]/text()")[0].replace(' ', '')
    book_info = tree.xpath("//div[@id='intro']/p[1]/text()")[0].replace(' ', '')
    info = book_name + '
    ' + author + '
    ' + book_type + '
    ' + last_update_time + '
    ' + book_info + '
    '
    dd_list = tree.xpath("//div[@id='list']/dl/dd/a/@href")
    
    
    def func(a):
        res = requests.get(url=f"https://www.biquge.info/41_41486/{a}",headers=headers)
        res.encoding = 'utf-8'
        res = res.text
        tree1 = etree.HTML(res)
        content_data = tree1.xpath('//div[@id="content"]//text()')
        name = tree1.xpath('//*[@id="wrapper"]/div[4]/div/div[2]/h1/text()')[0]  #小说章节
        content = '    ' + '
    '.join(content_data).lstrip()  #小说内容
        return name,content 
    
    
    f = open(f'./{book_name}.txt','w',encoding='utf-8')
    def write_xs(lst):
        for i in lst:
            print(i[0])
            f.write(i[0]+'
    '+i[1])
        f.close()
    
    pool = Pool(100)  #线程池
    lst = pool.map(func,dd_list)  #线程池访问页面
    write_xs(lst)  #调用写的函数
    
    
    print(time.time()-start)
    print('结束')
    import requests
    from lxml import etree
    from multiprocessing.dummy import Pool # 线程
    import time

    start = time.time()

    url = "https://www.biquge.info/41_41486/"
    headers = {
    "user-agent": "Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/73.0.3683.86 Safari/537.36"
    }
    page = requests.get(url=url, headers=headers)
    page.encoding = 'utf-8'
    page_text = page.text
    tree = etree.HTML(page_text)
    book_name = tree.xpath("//div[@id='info']/h1/text()")[0]
    author = tree.xpath("//div[@id='info']/p[1]/text()")[0].replace(' ', '').strip()
    book_type = tree.xpath("//div[@id='info']/p[2]/text()")[0].replace(' ', '')
    last_update_time = tree.xpath("//div[@id='info']/p[3]/text()")[0].replace(' ', '')
    book_info = tree.xpath("//div[@id='intro']/p[1]/text()")[0].replace(' ', '')
    info = book_name + ' ' + author + ' ' + book_type + ' ' + last_update_time + ' ' + book_info + ' '
    dd_list = tree.xpath("//div[@id='list']/dl/dd/a/@href")


    def func(a):
    res = requests.get(url=f"https://www.biquge.info/41_41486/{a}",headers=headers)
    res.encoding = 'utf-8'
    res = res.text
    tree1 = etree.HTML(res)
    content_data = tree1.xpath('//div[@id="content"]//text()')
    name = tree1.xpath('//*[@id="wrapper"]/div[4]/div/div[2]/h1/text()')[0]
    content = ' ' + ' '.join(content_data).lstrip()
    return name,content


    f = open(f'./{book_name}.txt','w',encoding='utf-8')
    def write_xs(lst):
    for i in lst:
    print(i[0])
    f.write(i[0]+' '+i[1])
    f.close()

    pool = Pool(100)
    lst = pool.map(func,dd_list)
    write_xs(lst)


    print(time.time()-start)
    print('结束')
  • 相关阅读:
    关于区间 $mex$ 的几种做法
    Selenium+TestNG+Maven+Jenkins+SVN(转载)
    Selenium+Java的TestNG测试报告优化
    使用Log4J收集日志
    Selenium+TestNG+Maven(2)
    自动化测试用例设计实例
    自动化测试用例的编写规则
    selenium 总结篇,常见方法和页面元素的操作
    Selenium Webdriver元素定位的八种常用方式(转载)
    selenium+java利用AutoIT实现文件上传
  • 原文地址:https://www.cnblogs.com/zhangchen-sx/p/10877343.html
Copyright © 2011-2022 走看看