zoukankan      html  css  js  c++  java
  • 模拟登陆

    1. github

    import requests
    from lxml import etree
    
    
    class Login(object):
        def __init__(self):
            self.headers = {
                'Origin': 'https://github.com',
                'Referer': 'https: // github.com /',
                'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/57.0.2987.133 Safari/537.36',
                'Host': 'github.com'
            }
            # 获取登录页面与token
            self.login_url = 'https://github.com/login'
            # 登录页面
            self.login_post_url = 'https://github.com/session'
            self.session=requests.session()
    
        # 获取token
        def get_token(self):
            headers = {
                'Host': 'github.com',
                'User-Agent': 'Mozilla/5.0 (Windows NT 6.1; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/72.0.3626.119 Safari/537.36',
    
            }
            res_text = self.session.get(url=self.login_url, headers=headers).text
            tree = etree.HTML(res_text)
            token = tree.xpath('//div//input[2]/@value')[0]
            return token
    
        # 模拟登录
        def login(self,username, password):
            token=self.get_token()
            print(token)
            post_data = {
                "login": username,
                "password":password,
                "commit": "Sign in",
                "utf8": "",
                "authenticity_token": token
             }
            res = self.session.post(url=self.login_post_url, data=post_data)
            print(res.status_code)
            if res.status_code == 200:
                self.get_email_page()
                with open("github.html", "wb") as f:
                    f.write(res.content)
        #获取个人页
        def get_email_page(self):
            print('获取个人页')
            email_data = self.session.get('https://github.com/tjp40922').text
            with open('xxxxxx.html', 'w', encoding='utf8') as f:
                f.write(email_data)
    
    
    
    if __name__ == '__main__':
        login = Login()
        username = input('请输入用户名')
        password = input('请输入密码')
        login.login(username, password)

      注意点:

        1.要获取token,而且要一一对应

        2.headers,两次请求的请求头是不一样的,否则会报错,422状态码

    2.scrapy模拟登陆人人

    # -*- coding: utf-8 -*-
    import scrapy
    from scrapy.http import Request,FormRequest
    
    class LoginrrSpider(scrapy.Spider):
        name = 'loginrr'
        allowed_domains = ['renren.com']
        start_urls = ['http://renren.com/']
        headers = {
            'User-Agent':
                'Mozilla/5.0 (Windows NT 10.0; WOW64) AppleWebKit/537.36 (KHTML, like Gecko) Chrome/63.0.3239.132 Safari/537.36'
        }
    
        # 这里我们先编写start_requests方法(会首先执行)
        def start_requests(self):
            # 首先访问一遍登录页面(在这里我们可以监测有无验证码,若有则需要添加验证码字段,若无则不需要)
            # 然后回调parse方法
            # 设置meta参数,将cookiejar设置为1,表示cookjar开启
            return [Request("http://www.renren.com/PLogin.do", meta={"cookiejar": 1}, callback=self.parse)]
    
        def parse(self, response):
            # 需要发送的表单,这次没有验证码
            data = {
                # 这里输入你们自己的账号密码
                'email': 'xxxxxx',
                'password': 'xxxxxxxx'
            }
            print("正在登陆...")
            # 通过 FormRequest.from_response方法来进行登录
            return [FormRequest.from_response(response,
                                              # 设置cookie信息
                                              meta={"cookiejar": response.meta["cookiejar"]},
                                              headers=self.headers,
                                              formdata=data,
                                              # 回调到next方法
                                              callback=self.next,
                                              )]
    
        def next(self, response):
            # 打印登录成功后的界面源代码
            print(response.body)

     3.scrapy模拟登录豆瓣

    import scrapy
    from faker import Factory
    

    #伪造ua的 f
    = Factory.create() class MailSpider(scrapy.Spider): name = 'douban-mail' allowed_domains = ['accounts.douban.com', 'douban.com'] start_urls = [ 'https://www.douban.com/' ] headers = { 'Accept': 'text/html,application/xhtml+xml,application/xml;q=0.9,*/*;q=0.8', 'Accept-Encoding': 'gzip, deflate, br', 'Accept-Language': 'zh-CN,zh;q=0.8,en-US;q=0.5,en;q=0.3', 'Connection': 'keep-alive', 'Host': 'accounts.douban.com', 'User-Agent': f.user_agent() } formdata = { 'form_email': '您的账号', 'form_password': '您的密码', # 'captcha-solution': '', # 'captcha-id': '', 'login': '登录', 'redir': 'https://www.douban.com/', 'source': 'None' } def start_requests(self): return [scrapy.Request(url='https://www.douban.com/accounts/login', headers=self.headers, meta={'cookiejar': 1}, callback=self.parse_login)] def parse_login(self, response): # 如果有验证码要人为处理 if 'captcha_image' in response.body: print 'Copy the link:' link = response.xpath('//img[@class="captcha_image"]/@src').extract()[0] print link captcha_solution = raw_input('captcha-solution:') captcha_id = urlparse.parse_qs(urlparse.urlparse(link).query, True)['id'] self.formdata['captcha-solution'] = captcha_solution self.formdata['captcha-id'] = captcha_id return [scrapy.FormRequest.from_response(response, formdata=self.formdata, headers=self.headers, meta={'cookiejar': response.meta['cookiejar']}, callback=self.after_login )] def after_login(self, response): print response.status self.headers['Host'] = "www.douban.com" return scrapy.Request(url='https://www.douban.com/doumail/', meta={'cookiejar': response.meta['cookiejar']}, headers=self.headers, callback=self.parse_mail) def parse_mail(self, response): print response.status for item in response.xpath('//div[@class="doumail-list"]/ul/li'): mail = DoubanMailItem() mail['sender_time'] = item.xpath('div[2]/div/span[1]/text()').extract()[0] mail['sender_from'] = item.xpath('div[2]/div/span[2]/text()').extract()[0] mail['url'] = item.xpath('div[2]/p/a/@href').extract()[0] mail['title'] = item.xpath('div[2]/p/a/text()').extract()[0] print mail yield mail
  • 相关阅读:
    js中连续赋值
    PHP文件上传漏洞原理以及如何防御详解 【转】
    奇虎360PHP工程师2019年面试题和答案解析【转】
    学习ES6碰到的单词
    Proxy(代理)例子
    for of 迭代器 生成器 数组的一些杂乱记录
    async和await
    Promise.call
    ES6中Promise.race
    Promise 和.then
  • 原文地址:https://www.cnblogs.com/tjp40922/p/10621522.html
Copyright © 2011-2022 走看看