zoukankan      html  css  js  c++  java
  • Python 爬虫实例(7)—— 爬取 新浪军事新闻

    我们打开新浪新闻,看到页面如下,首先去爬取一级 url,图片中蓝色圆圈部分

                             

    第二zh张图片,显示需要分页,

               

    源代码:

    # coding:utf-8
    
    import json
    import redis
    import time
    import requests
    session = requests.session()
    import logging.handlers
    import pickle
    import sys
    import re
    import datetime
    from bs4 import BeautifulSoup
    
    
    
    import sys
    reload(sys)
    sys.setdefaultencoding('utf8')
    
    import datetime
    # 生成一年的日期
    def dateRange(start, end, step=1, format="%Y-%m-%d"):
        strptime, strftime = datetime.datetime.strptime, datetime.datetime.strftime
        days = (strptime(end, format) - strptime(start, format)).days
        return [strftime(strptime(start, format) + datetime.timedelta(i), format) for i in xrange(0, days, step)]
    
    
    
    
    def spider():
    
        date_list = dateRange("2017-01-01", "2018-01-06")[::-1]
        print date_list
        for date in date_list:
            for page in range(1,5):
                #组合url
                url = "http://roll.mil.news.sina.com.cn/col/zgjq/" + str(date)+"_"+ str(page) +".shtml"
                # 伪装请求头
                headers = {
    
                    "Host":"roll.mil.news.sina.com.cn",
                  
                    "Cache-Control":"max-age=0",
                    "Upgrade-Insecure-Requests":"1",
                    "User-Agent":"Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/60.0.3112.113 Safari/537.36",
                    "Accept":"text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,*/*;q=0.8",
                  
                    "Accept-Encoding":"gzip, deflate",
                    "Accept-Language":"zh-CN,zh;q=0.8",
                    "If-Modified-Since":"Sat, 06 Jan 2018 09:57:24 GMT",
    
                }
    
                result = session.get(url=url,headers=headers).content
                #编码格式是 gb2312,使用BeautifulSoup解决编码格式
                soup = BeautifulSoup(result,'html.parser')
                #找到新闻列表
                result_div = soup.find_all('div',attrs={"class":"fixList"})[0]
                #去下换行
                result_replace = str(result_div).replace('
    ','').replace('
    ','').replace('	','')
                #正则匹配信息
                result_list = re.findall('<li>(.*?)</li>',result_replace)
    
                for i in result_list:
                    #匹配出来新闻 url, name,time
    
                    news_url = re.findall('<a href="(.*?)" target=',i)[0]
                    news_name = re.findall('target="_blank">(.*?)</a>',i)[0]
                    news_time = re.findall('<span class="time">((.*?))</span>',i)[0]
    
                    print news_url
                    print news_name
                    print news_time
    
    
    
    
    
    
    spider()
  • 相关阅读:
    Java学习之路(一)——JDK的下载与安装
    无法将“add-migration”项识别为 cmdlet、函数、脚本文件或可运行程序的名称。解决方案
    VS2019只能提示选中首选项的快捷键切换
    Visual Studio空格变成点的快捷键切换
    Visual Studio 2019 打开即时窗口
    完全卸载node.js
    安装node.js和vue
    在后台启动Redis
    mysql的数据库优化方案
    hadoop
  • 原文地址:https://www.cnblogs.com/xuchunlin/p/8215064.html
Copyright © 2011-2022 走看看