zoukankan      html  css  js  c++  java
  • python爬虫学习(2)__抓取糗百段子,与存入mysql数据库

    import pymysql
    import requests
    from bs4 import BeautifulSoup
    #pymysql链接数据库 conn
    =pymysql.connect(host='127.0.1',unix_socket='/tmp/mysql.sock',user='root',passwd='19950311',db='mysql') cur=conn.cursor() cur.execute("USE scraping")
    #存储段子标题,内容
    def store(title,content): cur.execute("insert into pages(title,content) values("%s","%s")",(title,content)) cur.connection.commit() global links class QiuShi(object): def __init__(self,start_url): self.url=start_url def crawing(self): try: html=requests.get(self.url,'lxml') return html.content except ConnectionError as e: return '' def extract(self,htmlContent): if len(htmlContent)>0: bsobj=BeautifulSoup(htmlContent,'lxml') #print bsobj jokes=bsobj.findAll('div',{'class':'article block untagged mb15'}) for j in jokes: text=j.find('h2').text content=j.find('div',{'class':'content'}).string if text != None and content != None: # print text,content,数据库编码为utf-8 store(text.encode('utf-8'),content.encode('utf-8')) print text.encode('utf-8'),content.encode('utf-8') print '------------------------------------------------------------------------------' else: print '' def main(self): text=self.crawing() self.extract(text) try: qiushi=QiuShi('http://www.qiushibaike.com/') qiushi.main() finally:
    #关闭cursor,connection cur.close() conn.close()
  • 相关阅读:
    Visual SVN1.5以上版本下载及破解过程
    C#线程系列讲座(2):Thread类的应用
    a标签的href 和onclick
    Windows 服务快捷启动命令
    iframe 跨域自动适应高度
    修正认知:string、Empty和null三者的正确区别

    线索二叉树
    最大连续子序列
    寻找最近点对
  • 原文地址:https://www.cnblogs.com/yunwuzhan/p/5765963.html
Copyright © 2011-2022 走看看