This time I will bring you a detailed explanation of the php record searchindexengine crawling process method, what are theprecautionsfor php record search engine crawling process, the following is a practical case, let’s take a look take a look.
The following is the complete code:
//记录搜索引擎爬行记录 $searchbot = get_naps_bot(); if ($searchbot) { $tlc_thispage = addslashes($_SERVER['HTTP_USER_AGENT']); $url = $_SERVER['HTTP_REFERER']; $file = WEB_PATH.'robotslogs.txt'; $date = date('Y-m-d H:i:s'); $data = fopen($file,'a'); fwrite($data,"Time:$date robot:$searchbot URL:$tlc_thispage/r/n"); fclose($data); }
WEB_PATH is the root directory path of define under index.PHP, which means that the robotslogs.txt file is placed in the root directory.
Get the spider crawling record throughget_naps_bot()
, then process it through addslashes, and store the data in thevariable$tlc_thispage.
fopen opens the robotslogs.txt file, writes the data through the function fwrite, and closes it through the function fclose.
Because I felt it was unnecessary, I deleted the code on my website, so there are no examples of the effect.
PS: php code to obtain the crawling records of each search spider
Supports the following search engines: Baidu, Google, Bing, Yahoo, Soso , Sogou, Yodao crawling website records!
Code:
I believe you have mastered the method after reading the case in this article. For more exciting information, please pay attention to other related articles on the PHP Chinese website!
Recommended reading:
What are the methods for php to read local json files
What are the methods for php to output json objects The value of
The above is the detailed content of Detailed explanation of the method of recording search engine crawling process in PHP. For more information, please follow other related articles on the PHP Chinese website!