1、UA随机
2、操作简单方便,直接getRank(关键词,域名)就可以了
3、编码转化。编码方面应该没啥问题了。
4、结果丰富。不仅有排名,还有搜索结果的title,URL,快照时间,符合SEO需求
缺点:
单线程,速度慢
# @param python查询百度SEO信息
# @author 编程之家 jb51.cc|www.jb51.cc
#coding=utf-8
import requests
import BeautifulSoup
import re
import random
def decodeAnyWord(w):
try:
w.decode('utf-8')
except:
w = w.decode('gb2312')
else:
w = w.decode('utf-8')
return w
def createURL(checkWord): #create baidu URL with search words
checkWord = checkWord.strip()
checkWord = checkWord.replace(' ','+').replace('\n','')
baiduURL = 'http://www.baidu.com/s?wd=%s&rn=100' % checkWord
return baiduURL
def getContent(baiduURL): #get the content of the serp
uaList = ['Mozilla/4.0+(compatible;+MSIE+6.0;+Windows+NT+5.1;+SV1;+.NET+CLR+1.1.4322;+TencentTraveler)','Mozilla/4.0+(compatible;+MSIE+6.0;+Windows+NT+5.1;+SV1;+.NET+CLR+2.0.50727;+.NET+CLR+3.0.4506.2152;+.NET+CLR+3.5.30729)','Mozilla/5.0+(Windows+NT+5.1)+AppleWebKit/537.1+(KHTML,+like+Gecko)+Chrome/21.0.1180.89+Safari/537.1','Mozilla/4.0+(compatible;+MSIE+6.0;+Windows+NT+5.1;+SV1)','Mozilla/5.0+(Windows+NT+6.1;+rv:11.0)+Gecko/20100101+Firefox/11.0','Mozilla/4.0+(compatible;+MSIE+8.0;+Windows+NT+5.1;+Trident/4.0;+SV1)','Mozilla/4.0+(compatible;+MSIE+8.0;+Windows+NT+5.1;+Trident/4.0;+GTB7.1;+.NET+CLR+2.0.50727)','Mozilla/4.0+(compatible;+MSIE+8.0;+Windows+NT+5.1;+Trident/4.0;+KB974489)']
headers = {'User-Agent': random.choice(uaList)}
ipList = ['202.43.188.13:8080','80.243.185.168:1177','218.108.85.59:81']
proxies = {'http': 'http://%s' % random.choice(ipList)}
r = requests.get(baiduURL,headers = headers,proxies = proxies)
return r.content
def getLastURL(rawurl): #get final URL while there're redirects
r = requests.get(rawurl)
return r.url
def getAtext(atext): #get the text with <a> and </a>
pat = re.compile(r'<a .*?>(.*?)</a>')
match = pat.findall(atext)
pureText = match[0].replace('<em>','').replace('</em>','')
return pureText
def getCacheDate(t): #get the date of cache
pat = re.compile(r'<span class="g">.*?(\d{4}-\d{1,2}-\d{1,2}) </span>')
match = pat.findall(t)
cacheDate = match[0]
return cacheDate
def getRank(checkWord,domain): #main line
checkWord = checkWord.replace('\n','')
checkWord = decodeAnyWord(checkWord)
baiduURL = createURL(checkWord)
cont = getContent(baiduURL)
soup = BeautifulSoup.BeautifulSoup(cont)
results = soup.findAll('table',{'class': 'result'}) #find all results in this page
for result in results:
checkData = unicode(result.find('span',{'class': 'g'}))
if re.compile(r'^[^/]*%s.*?' %domain).match(checkData): #改正则
nowRank = result['id'] #get the rank if match the domain info
resLink = result.find('h3').a
resURL = resLink['href']
domainURL = getLastURL(resURL) #get the target URL
resTitle = getAtext(unicode(resLink)) #get the title of the target page
rescache = result.find('span',{'class': 'g'})
cacheDate = getCacheDate(unicode(rescache)) #get the cache date of the target page
res = u'%s,第%s名,%s,%s' % (checkWord,nowRank,resTitle,cacheDate,domainURL)
return res.encode('gb2312')
break
else:
return '>100'
domain = 'www.douban.com' #set the domain which you want to search.
f = open('r.txt')
for w in f.readlines():
print getRank(w,domain)
f.close()
# End www.jb51.cc