Let’s talk about our school’s website first:
http://jwxt.sdu.edu.cn:7777/zhxt_bks/zhxt_bks.html
To check scores, you need to log in and then display The results of each subject are displayed, but only the results are displayed without the grade points, which is the weighted average score.
Obviously calculating grade points manually is a very troublesome thing. So we can use Python to make a crawler to solve this problem.
1. On the eve of the decisive battle
Let’s prepare a tool first: HttpFox plug-in.
This is an http protocol analysis plug-in that analyzes page request and response time, content, and COOKIE used by the browser.
Take me as an example, just install it on Firefox, the effect is as shown:
Okay View the corresponding information very intuitively.
Click start to start detection, click stop to pause detection, and click clear to clear the content.
Generally before use, click stop to pause, and then click clear to clear the screen to ensure that you see the data obtained by accessing the current page.
2. Go deep behind enemy lines
Let’s go to Shandong University’s score query website to see what information is sent when logging in.
First go to the login page, open httpfox, after clearing, click start to turn on the detection:
After entering the personal information, make sure httpfox is turned on. Then click OK to submit the information and log in.
You can see at this time that httpfox has detected three pieces of information:
At this time, click the stop button to ensure that what is captured is the feedback after visiting the page. Data so that we can simulate login when doing crawlers.
3. Pao Ding Jie Niu
At first glance, we got three data, two are GET and one is POST, but what are they? What and how it should be used, we still have no idea.
So, we need to check the captured content one by one.
Look at the POST information first:
Since it is the POST information, we can just look at the PostData.
You can see that there are two POST data, studid and pwd.
And it can be seen from the Redirect to of Type that after the POST is completed, it jumps to the bks_login2.loginmessage page.
It can be seen that this data is the form data submitted after clicking OK.
Click on the cookie label to see the cookie information:
Yes, an ACCOUNT cookie was received and will be automatically destroyed after the session ends.
So what information did you receive after submitting?
Let’s take a look at the next two GET data.
Look at the first one first. We click on the content tag to view the received content. Do you feel like eating it alive? -The HTML source code is undoubtedly exposed:
It seems that this is just the html source code of the page. Click on the cookie to view the cookie-related information:
Aha, it turns out that the content of the html page was received only after the cookie information was sent.
Let’s take a look at the last received message:
After a rough look, it should be just a css file called style.css, which doesn’t mean much to us. big effect.
4. Calmly respond
Now that we know what data we sent to the server and what data we received, the basic process is as follows:
First, we POST the student ID and password--->Then return the cookie value
Then send the cookie to the server--->Return the page information.
Get the data from the grades page, use regular expressions to extract the grades and credits separately and calculate the weighted average.
OK, it looks like a very simple sample paper. Then let’s try it out.
But before the experiment, there is still an unresolved problem, which is where is the POST data sent?
Look at the original page again:
It is obviously implemented using an html framework, that is to say, what we see in the address bar The address is not the address to submit the form on the right.
So how can I get the real address-. -Right click to view page source code:
Yes, that’s right, the one with name="w_right" is the login page we want.
The original address of the website is:
http://jwxt.sdu.edu.cn:7777/zhxt_bks/zhxt_bks.html
So, the real form submission The address should be:
http://jwxt.sdu.edu.cn:7777/zhxt_bks/xk_login.html
After entering it, it is true:
It’s actually the course selection system of Tsinghua University. . . My guess is that our school was too lazy to create a page, so we just borrowed it. . As a result, the title was not even changed. . .
But this page is still not the page we need, because the page our POST data is submitted to should be the page submitted in the ACTION of the form.
In other words, we need to check the source code to know where the POST data is sent:
Well, visually this is the address where the POST data is submitted. .
Arrange it into the address bar. The complete address should be as follows:
http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login
(The method of obtaining it is very simple. Just click on the link directly in Firefox browser to see the address of the link)
5. A little test of my skills
The next task is to use python to simulate sending a POST data and get the returned cookie value.
Regarding the operation of cookies, you can read this blog post:
http://blog.csdn.net/wxg694175346/article/details/8925978
We first prepare a POST data, then prepare a cookie to receive, and then write the source code as follows:
# -*- coding: utf-8 -*- #--------------------------------------- # 程序:山东大学爬虫 # 版本:0.1 # 作者:why # 日期:2013-07-12 # 语言:Python 2.7 # 操作:输入学号和密码 # 功能:输出成绩的加权平均值也就是绩点 #--------------------------------------- import urllib import urllib2 import cookielib cookie = cookielib.CookieJar() opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cookie)) #需要POST的数据# postdata=urllib.urlencode({ 'stuid':'201100300428', 'pwd':'921030' }) #自定义一个请求# req = urllib2.Request( url = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login', data = postdata ) #访问该链接# result = opener.open(req) #打印返回的内容# print result.read()
After this, look at the effect of the operation:
ok, in this way, we will calculate that the simulated login is successful.
6. Change the situation
The next task is to use a crawler to obtain the students’ scores.
Let’s look at the source website again.
After opening HTTPFOX, click to view the results and find that the following data was captured:
##Click on the first GET data and check the content. You can find that the Content is the content of the obtained results.
## so you can get The link to view the results is as follows:
http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bkscjcx.curscopere
7. Everything is ready
Now everything is ready, so just apply the link to the crawler and see if you can see the results page.
As you can see from httpfox, we have to send a cookie to return the score information, so we use python to simulate sending a cookie to request the score information:
# -*- coding: utf-8 -*- #--------------------------------------- # 程序:山东大学爬虫 # 版本:0.1 # 作者:why # 日期:2013-07-12 # 语言:Python 2.7 # 操作:输入学号和密码 # 功能:输出成绩的加权平均值也就是绩点 #--------------------------------------- import urllib import urllib2 import cookielib #初始化一个CookieJar来处理Cookie的信息# cookie = cookielib.CookieJar() #创建一个新的opener来使用我们的CookieJar# opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(cookie)) #需要POST的数据# postdata=urllib.urlencode({ 'stuid':'201100300428', 'pwd':'921030' }) #自定义一个请求# req = urllib2.Request( url = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login', data = postdata ) #访问该链接# result = opener.open(req) #打印返回的内容# print result.read() #打印cookie的值 for item in cookie: print 'Cookie:Name = '+item.name print 'Cookie:Value = '+item.value #访问该链接# result = opener.open('http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bkscjcx.curscopre') #打印返回的内容# print result.read()
Since there is no problem, use regular expressions to convert the data Just process it a little and take out the credits and corresponding scores.
8. At your fingertips
Such a large amount of html source code is obviously not conducive to our processing. Next, we will use regular expressions to extract the necessary data.
For tutorials on regular expressions, you can read this blog post:
http://blog.csdn.net/wxg694175346/article/details/8929576
Let’s take a look at the source code of the score:
既然如此,用正则表达式就易如反掌了。 我们将代码稍稍整理一下,然后用正则来取出数据: 水平有限,,正则是有点丑,。运行的效果如图: ok,接下来的只是数据的处理问题了。。 9.凯旋而归 完整的代码如下,至此一个完整的爬虫项目便完工了。 以上就是 [Python]网络爬虫(十):一个爬虫的诞生全过程(以山东大学绩点运算为例)的内容,更多相关内容请关注PHP中文网(m.sbmmt.com)!# -*- coding: utf-8 -*-
#---------------------------------------
# 程序:山东大学爬虫
# 版本:0.1
# 作者:why
# 日期:2013-07-12
# 语言:Python 2.7
# 操作:输入学号和密码
# 功能:输出成绩的加权平均值也就是绩点
#---------------------------------------
import urllib
import urllib2
import cookielib
import re
class SDU_Spider:
# 申明相关的属性
def __init__(self):
self.loginUrl = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login' # 登录的url
self.resultUrl = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bkscjcx.curscopre' # 显示成绩的url
self.cookieJar = cookielib.CookieJar() # 初始化一个CookieJar来处理Cookie的信息
self.postdata=urllib.urlencode({'stuid':'201100300428','pwd':'921030'}) # POST的数据
self.weights = [] #存储权重,也就是学分
self.points = [] #存储分数,也就是成绩
self.opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(self.cookieJar))
def sdu_init(self):
# 初始化链接并且获取cookie
myRequest = urllib2.Request(url = self.loginUrl,data = self.postdata) # 自定义一个请求
result = self.opener.open(myRequest) # 访问登录页面,获取到必须的cookie的值
result = self.opener.open(self.resultUrl) # 访问成绩页面,获得成绩的数据
# 打印返回的内容
# print result.read()
self.deal_data(result.read().decode('gbk'))
self.print_data(self.weights);
self.print_data(self.points);
# 将内容从页面代码中抠出来
def deal_data(self,myPage):
myItems = re.findall('<TR>.*?<p.*?<p.*?<p.*?<p.*?<p.*?>(.*?)</p>.*?<p.*?<p.*?>(.*?)</p>.*?</TR>',myPage,re.S) #获取到学分
for item in myItems:
self.weights.append(item[0].encode('gbk'))
self.points.append(item[1].encode('gbk'))
# 将内容从页面代码中抠出来
def print_data(self,items):
for item in items:
print item
#调用
mySpider = SDU_Spider()
mySpider.sdu_init()
# -*- coding: utf-8 -*-
#---------------------------------------
# 程序:山东大学爬虫
# 版本:0.1
# 作者:why
# 日期:2013-07-12
# 语言:Python 2.7
# 操作:输入学号和密码
# 功能:输出成绩的加权平均值也就是绩点
#---------------------------------------
import urllib
import urllib2
import cookielib
import re
import string
class SDU_Spider:
# 申明相关的属性
def __init__(self):
self.loginUrl = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bks_login2.login' # 登录的url
self.resultUrl = 'http://jwxt.sdu.edu.cn:7777/pls/wwwbks/bkscjcx.curscopre' # 显示成绩的url
self.cookieJar = cookielib.CookieJar() # 初始化一个CookieJar来处理Cookie的信息
self.postdata=urllib.urlencode({'stuid':'201100300428','pwd':'921030'}) # POST的数据
self.weights = [] #存储权重,也就是学分
self.points = [] #存储分数,也就是成绩
self.opener = urllib2.build_opener(urllib2.HTTPCookieProcessor(self.cookieJar))
def sdu_init(self):
# 初始化链接并且获取cookie
myRequest = urllib2.Request(url = self.loginUrl,data = self.postdata) # 自定义一个请求
result = self.opener.open(myRequest) # 访问登录页面,获取到必须的cookie的值
result = self.opener.open(self.resultUrl) # 访问成绩页面,获得成绩的数据
# 打印返回的内容
# print result.read()
self.deal_data(result.read().decode('gbk'))
self.calculate_date();
# 将内容从页面代码中抠出来
def deal_data(self,myPage):
myItems = re.findall('<TR>.*?<p.*?<p.*?<p.*?<p.*?<p.*?>(.*?)</p>.*?<p.*?<p.*?>(.*?)</p>.*?</TR>',myPage,re.S) #获取到学分
for item in myItems:
self.weights.append(item[0].encode('gbk'))
self.points.append(item[1].encode('gbk'))
#计算绩点,如果成绩还没出来,或者成绩是优秀良好,就不运算该成绩
def calculate_date(self):
point = 0.0
weight = 0.0
for i in range(len(self.points)):
if(self.points[i].isdigit()):
point += string.atof(self.points[i])*string.atof(self.weights[i])
weight += string.atof(self.weights[i])
print point/weight
#调用
mySpider = SDU_Spider()
mySpider.sdu_init()