zoukankan      html  css  js  c++  java
  • python爬微博

    # -*- coding: utf-8 -*-
    
    import urllib.request
    import json
    
    #定义要爬取的微博大V的微博ID
    id='3924739974'
    
    #设置代理IP
    proxy_addr="122.241.72.191:808"
    
    #定义页面打开函数
    def use_proxy(url,proxy_addr):
        req=urllib.request.Request(url)
        req.add_header("User-Agent","Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/49.0.2623.221 Safari/537.36 SE 2.X MetaSr 1.0")
        proxy=urllib.request.ProxyHandler({'http':proxy_addr})
        opener=urllib.request.build_opener(proxy,urllib.request.HTTPHandler)
        urllib.request.install_opener(opener)
        data=urllib.request.urlopen(req).read().decode('utf-8','ignore')
        return data
    
    #获取微博主页的containerid,爬取微博内容时需要此id
    def get_containerid(url):
        data=use_proxy(url,proxy_addr)
        content=json.loads(data).get('data')
        for data in content.get('tabsInfo').get('tabs'):
            if(data.get('tab_type')=='weibo'):
                containerid=data.get('containerid')
        return containerid
    
    #获取微博大V账号的用户基本信息,如:微博昵称、微博地址、微博头像、关注人数、粉丝数、性别、等级等
    def get_userInfo(id):
        url='https://m.weibo.cn/api/container/getIndex?type=uid&value='+id
        data=use_proxy(url,proxy_addr)
        content=json.loads(data).get('data')
        profile_image_url=content.get('userInfo').get('profile_image_url')
        description=content.get('userInfo').get('description')
        profile_url=content.get('userInfo').get('profile_url')
        verified=content.get('userInfo').get('verified')
        guanzhu=content.get('userInfo').get('follow_count')
        name=content.get('userInfo').get('screen_name')
        fensi=content.get('userInfo').get('followers_count')
        gender=content.get('userInfo').get('gender')
        urank=content.get('userInfo').get('urank')
        print("微博昵称:"+name+"
    "+"微博主页地址:"+profile_url+"
    "+"微博头像地址:"+profile_image_url+"
    "+"是否认证:"+str(verified)+"
    "+"微博说明:"+description+"
    "+"关注人数:"+str(guanzhu)+"
    "+"粉丝数:"+str(fensi)+"
    "+"性别:"+gender+"
    "+"微博等级:"+str(urank)+"
    ")
    
    
    #获取微博内容信息,并保存到文本中,内容包括:每条微博的内容、微博详情页面地址、点赞数、评论数、转发数等
    def get_weibo(id,file):
        i=1
    
        url = 'https://m.weibo.cn/api/container/getIndex?type=uid&value=' + id
        weibo_url = 'https://m.weibo.cn/api/container/getIndex?type=uid&value=' + id + '&containerid=' + get_containerid(
            url) + '&page=' + str(i)
        try:
            data = use_proxy(weibo_url, proxy_addr)
            content = json.loads(data).get('data')
            cards = content.get('cards')
            print(content)
            '''
            if(len(cards)>0):
                for j in range(len(cards)):
                    print("-----正在爬取第"+str(i)+"页,第"+str(j)+"条微博------")
                    card_type=cards[j].get('card_type')
                    if(card_type==9):
                        mblog=cards[j].get('mblog')
                        attitudes_count=mblog.get('attitudes_count')
                        comments_count=mblog.get('comments_count')
                        created_at=mblog.get('created_at')
                        reposts_count=mblog.get('reposts_count')
                        scheme=cards[j].get('scheme')
                        text=mblog.get('text')
                        with open(file,'a',encoding='utf-8') as fh:
                            fh.write("----第"+str(i)+"页,第"+str(j)+"条微博----"+"
    ")
                            fh.write("微博地址:"+str(scheme)+"
    "+"发布时间:"+str(created_at)+"
    "+"微博内容:"+text+"
    "+"点赞数:"+str(attitudes_count)+"
    "+"评论数:"+str(comments_count)+"
    "+"转发数:"+str(reposts_count)+"
    ")
                i+=1
            else:
                break
            '''
        except Exception as e:
            print(e)
            pass
    
        '''
        while True:
            url='https://m.weibo.cn/api/container/getIndex?type=uid&value='+id
            weibo_url='https://m.weibo.cn/api/container/getIndex?type=uid&value='+id+'&containerid='+get_containerid(url)+'&page='+str(i)
            try:
                data=use_proxy(weibo_url,proxy_addr)
                content=json.loads(data).get('data')
                cards=content.get('cards')
                print(content)
    
                if(len(cards)>0):
                    for j in range(len(cards)):
                        print("-----正在爬取第"+str(i)+"页,第"+str(j)+"条微博------")
                        card_type=cards[j].get('card_type')
                        if(card_type==9):
                            mblog=cards[j].get('mblog')
                            attitudes_count=mblog.get('attitudes_count')
                            comments_count=mblog.get('comments_count')
                            created_at=mblog.get('created_at')
                            reposts_count=mblog.get('reposts_count')
                            scheme=cards[j].get('scheme')
                            text=mblog.get('text')
                            with open(file,'a',encoding='utf-8') as fh:
                                fh.write("----第"+str(i)+"页,第"+str(j)+"条微博----"+"
    ")
                                fh.write("微博地址:"+str(scheme)+"
    "+"发布时间:"+str(created_at)+"
    "+"微博内容:"+text+"
    "+"点赞数:"+str(attitudes_count)+"
    "+"评论数:"+str(comments_count)+"
    "+"转发数:"+str(reposts_count)+"
    ")
                    i+=1
                else:
                    break
    
            except Exception as e:
                print(e)
                pass
        '''
    if __name__=="__main__":
        file=id+".txt"
        get_userInfo(id)
        get_weibo(id,file)
    今日事今日毕
  • 相关阅读:
    BNU校赛
    Latest Common Ancestor
    Codeforces Round #482 (Div. 2)
    Persistent Line Segment Tree
    2018HNCCPC(Onsite)
    2018HNCCPC
    2017 ACM Jordanian Collegiate Programming Contest
    Codeforces Round #480 (Div. 2)
    负载均衡SLB
    windows下的端口监听、程序端口查找命令
  • 原文地址:https://www.cnblogs.com/gjack/p/8351709.html
Copyright © 2011-2022 走看看