zoukankan      html  css  js  c++  java
  • py3+requests+urllib+bs4+threading,爬取斗图图片

    实现原理及思路请参考我的另外几篇爬虫实践博客

    py3+urllib+bs4+反爬,20+行代码教你爬取豆瓣妹子图:http://www.cnblogs.com/UncleYong/p/6892688.html
    py3+requests+json+xlwt,爬取拉勾招聘信息:http://www.cnblogs.com/UncleYong/p/6960044.html
    py3+urllib+re,轻轻松松爬取双色球最近100期中奖号码:http://www.cnblogs.com/UncleYong/p/6958242.html

    实现代码如下:

    #-*- coding:utf-8 -*-
    import requests, threading, time
    from lxml import etree
    from bs4 import BeautifulSoup
    
    # 获取源码
    def get_html(url):
        # url = 'http://www.doutula.com/article/list/?page=1'
        headers = {'user-agent':'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/58.0.3029.110 Safari/537.36'}
        request = requests.get(url=url, headers=headers) # 网址发送get请求
        response = request.content.decode('utf-8') # 获取源码
        # print(response)
        return response
    
    # 匹配图片url
    def get_img_html(html):
        # soup = BeautifulSoup(html,'html.parser')
        soup = BeautifulSoup(html,'lxml') # 解析网页
        all_a = soup.find_all('a',class_='list-group-item') # 获取a标签,如果有class或id来命名,一定要加上名字
        # class="list-group-item"是a标签的名字
        # <a class="list-group-item" href="http://www.doutula.com/article/detail/7536783">
        # print(type(all_a)) # <class 'bs4.element.ResultSet'>
        # print(all_a)
        for i in all_a:
            # print(i['href'])
            img_html = get_html(i['href']) # 获取内页源码,i['href']表示获取属性值
            # print(img_html)
            get_img(img_html)
    # 获取图片url def get_img(html): # soup = etree.HTML(html) # 初始化源码 # items = soup.xpath('//div[@class="artile_des"]') # //表示某个目录下,从匹配选择的当前节点选择文档中的节点,而不考虑它们的位置。 # # []表示过滤条件 # for item in items: # imgurl_list = item.xpath('table/tbody/tr/td/a/img/@onerror') # # print(imgurl_list) # # start_save_img(imgurl_list) soup = BeautifulSoup(html, 'lxml') items = soup.find('div',class_='swiper-slide').find_all('div',class_='artile_des') # 不能写成这样:find_all后面不能跟find,因为find是找一个,find_all是找多个,从多个中找一个是不对的 # items = soup.find('div',class_='swiper-slide').find_all('div',class_='artile_des').find('img')['src'] # print(items) imgurl_list = [] for i in items: imgurl = i.find('img')['src'] # img标签下的src属性 # print(type(imgurl)) # <class 'str'> # print(imgurl) imgurl_list.append(imgurl) start_save_img(imgurl_list) # 这里是对每一组套图做多线程 # 下载图片 x = 1 def save_img(img_url): # global x # 全局变量 # x +=1 # img_url = img_url.split('=')[-1][1:-2].replace('jp','jpg') # 以=分割 # print('正在下载'+'http:'+img_url) # img_content = requests.get('http:'+img_url).content # with open('doutu/%s.jpg'%x, 'wb') as f:# urllib下的retrieve也可以下载 # f.write(img_content) global x # 全局变量 x +=1 print('正在下载:'+img_url) geshi = img_url.split('.')[-1] # 因为图片格式不一样,所以切片,把链接中图片后缀获取到,用于下面拼接文件名 img_content = requests.get(img_url).content with open('doutu/%s.%s'%(x,geshi), 'wb') as f: # urllib下的retrieve也可以下载 f.write(img_content) def start_save_img(imgurl_list): for i in imgurl_list: # print(i) th = threading.Thread(target=save_img,args=(i,)) # i后面加逗号表示args是一个元组 # target是可调用对象,是一个函数名,线程启动后执行, th.start() th.join()
    # 主函数 def main(): start_url = 'http://www.doutula.com/article/list/?page={}' for i in range(1,2): # print(start_url.format(i)) start_html = get_html(start_url.format(i)) get_img_html(start_html) # 获取内页图片的url if __name__ == '__main__': # 判断文件入口 start_time = time.time() main() end_time = time.time() print(start_time) print(end_time) print(end_time-start_time)

  • 相关阅读:
    Java Image Processing
    贝塞尔曲线开发的艺术
    Ubuntu中Hadoop环境搭建
    FIRST集合、FOLLOW集合、SELECT集合以及预测分析表地构造
    Linux环境下使用VSCode编译makefile文件的注意事项
    神经记忆模型
    深度学习推荐阅读的论文
    博客园无法发布文章解决办法
    计算机各个方向名校公开课
    软件过程基础
  • 原文地址:https://www.cnblogs.com/uncleyong/p/6973113.html
Copyright © 2011-2022 走看看