python爬虫CSDN文章抓取

来源:互联网 发布:ubuntu find 文件夹 编辑:程序博客网 时间:2024/06/10 09:38

CSDN原则上不让非人浏览访问,正常爬虫无法从这里爬取文章,需要进行模拟人为浏览器访问。


使用:输入带文章的CSDN链接自动生成正文的HTML,文件名为标题名


#!/usr/bin/env python# coding=utf-8##########################################> File Name: CSDN_article.py#> Author: nealgavin#> Mail: nealgavin@126.com #> Created Time: Tue 27 May 2014 03:42:54 PM CST#########################################import randomimport socketimport urllib2import urllibimport reimport stringimport BeautifulSoupimport sysimport cookielibERROR = {        '0':'Can not open the url,checck you net',        '1':'Creat download dir error',        '2':'The image links is empty',        '3':'Download faild',        '4':'Build soup error,the html is empty',        '5':'Can not save the image to your disk',    }class BrowserBase(object):     """模拟浏览器"""    def __init__(self):        socket.setdefaulttimeout(20)        self.HTML = ''        self.articleName = ''        self.link = ''    def speak(self,name,content):        print '[%s]%s' %(name,content)    def openurl(self,url):        """        打开网页        """        cookie_support= urllib2.HTTPCookieProcessor(cookielib.CookieJar())        self.opener = urllib2.build_opener(cookie_support,urllib2.HTTPHandler)        urllib2.install_opener(self.opener)        user_agents = [                    'Mozilla/5.0 (Windows; U; Windows NT 5.1; it; rv:1.8.1.11) Gecko/20071127 Firefox/2.0.0.11',                    'Opera/9.25 (Windows NT 5.1; U; en)',                    'Mozilla/4.0 (compatible; MSIE 6.0; Windows NT 5.1; SV1; .NET CLR 1.1.4322; .NET CLR 2.0.50727)',                    'Mozilla/5.0 (compatible; Konqueror/3.5; Linux) KHTML/3.5.5 (like Gecko) (Kubuntu)',                    'Mozilla/5.0 (X11; U; Linux i686; en-US; rv:1.8.0.12) Gecko/20070731 Ubuntu/dapper-security Firefox/1.5.0.12',                    'Lynx/2.8.5rel.1 libwww-FM/2.14 SSL-MM/1.4.1 GNUTLS/1.2.9',                    "Mozilla/5.0 (X11; Linux i686) AppleWebKit/535.7 (KHTML, like Gecko) Ubuntu/11.04 Chromium/16.0.912.77 Chrome/16.0.912.77 Safari/535.7",                    "Mozilla/5.0 (X11; Ubuntu; Linux i686; rv:10.0) Gecko/20100101 Firefox/10.0 ",                    ]                agent = random.choice(user_agents)        self.opener.addheaders = [("User-agent",agent),("Accept","*/*"),('Referer','http://www.google.com')]        try:            res = self.opener.open(url)            self.HTML = res.read()#            print res.read()        except Exception,e:            self.speak(str(e),url)            raise Exception        else:            return res    def OUT(self):        print self.HTML    def getArticleName(self,tags):        re_rules = r'<span class="(.+?)"><a href="(.+?)">(.+?)</a>'        p = re.compile(re_rules,re.DOTALL)        title = p.findall(str(tags))        self.chineseListOut(title)    def chineseListOut(self,tags):        title = []        for tag in tags:            for ele in tag:#                print '+',str(ele),'-'                title.append(ele.strip())        self.link = "http://blog.csdn.net" + title[1]        tle = title[2].split()        self.articleName = '-'.join(tle)    def buildArticleHTML(self):        self.HTML = str(self.HTML)        self.HTML = '<html><meta charset="utf-8"><body>'+self.HTML        self.HTML = self.HTML + '</body><html>'    def getMainArticle(self):        """get the main article of CSDN blog"""        soup = BeautifulSoup.BeautifulSoup(self.HTML)        tags_all = soup.findAll('div',{'class':'article_title'})        self.getArticleName(tags_all)        tags_all = soup.findAll('div',{'id':'article_content','class':'article_content'})        self.HTML = tags_all[0]        self.buildArticleHTML()        def saveArticleAsHTML(self):        filePath = self.articleName+'.html'        try:                        filePointer = open(filePath,'w+')        except:            print 'open error'        print 'path = ',filePath        filePointer.write(self.HTML)        filePointer.close()browser = BrowserBase()url = raw_input('Input the links of CSDN article you needed!\n')if url is None or len(url) == 0:    url = "http://blog.csdn.net/nealgavin/article/details/27110717"browser.openurl(url)browser.getMainArticle()browser.saveArticleAsHTML()


1 0
原创粉丝点击