Home>Article>Web Front-end> Example of node crawling data: Let’s talk about how to crawl novel chapters

Example of node crawling data: Let’s talk about how to crawl novel chapters

青灯夜游
青灯夜游 forward
2022-05-02 10:00:17 3658browse

How does node crawl data? The following article will share with you anodecrawler example and talk about how to use node to crawl novel chapters. I hope it will be helpful to everyone!

Example of node crawling data: Let’s talk about how to crawl novel chapters

I am going to useelectronto make a novel reading tool to practice. The first thing to solve is the data problem, that is, the text of the novel.

Here we are going to usenodejsto crawl the novel website, try to crawl the next novel, the data will not be stored in the database, first usetxtas text storage

For website requests innode, there arehttpandhttpslibraries, which containrequestrequests. method.

Example:

request = https.request(TestUrl, { encoding:'utf-8' }, (res)=>{ let chunks = '' res.on('data', (chunk)=>{ chunks += chunk }) res.on('end',function(){ console.log('请求结束'); }) })

But that’s it. It only accesses ahtmltext data, and cannot extract internal elements. Work (you can also get it regularly, but it’s too complicated).

I stored the accessed data through thefs.writeFilemethod. This is just the html of the entire web page

Example of node crawling data: Let’s talk about how to crawl novel chapters

But I What you want is the content in each chapter. In this way, you need to obtain the hyperlink of the chapter, form a hyperlink linked list and crawl it

Example of node crawling data: Let’s talk about how to crawl novel chapters

cheerio library

So, here we will introduce a js library,cheerio

##Official document: https://cheerio.js.org/

Chinese documentation: https://github.com/cheeriojs/cheerio/wiki/Chinese-README

In the documentation, you can use examples for debugging

Example of node crawling data: Let’s talk about how to crawl novel chapters


Use cheerio to parse HTML

When cheerio parses html, the method of obtaining dom nodes is similar to

jquery.

Find the dom node data you want based on the html of the book home page obtained previously

const fs = require('fs') const cheerio = require('cheerio'); // 引入读取方法 const { getFile, writeFun } = require('./requestNovel') let hasIndexPromise = getFile('./hasGetfile/index.html'); let bookArray = []; hasIndexPromise.then((res)=>{ let htmlstr = res; let $ = cheerio.load(htmlstr); $(".listmain dl dd a").map((index, item)=>{ let name = $(item).text(), href = 'https://www.shuquge.com/txt/147032/' + $(item).attr('href') if (index > 11){ bookArray.push({ name, href }) } }) // console.log(bookArray) writeFun('./hasGetfile/hrefList.txt', JSON.stringify(bookArray), 'w') })

Print the information

Example of node crawling data: Let’s talk about how to crawl novel chapters

Okay At the same time, this information is also stored

Example of node crawling data: Let’s talk about how to crawl novel chapters


Now that the number of chapters and links to chapters are available, you can obtain the content of the chapters.

Because batch crawling requires an IP proxy in the end, we are not ready yet. For the time being, we will write a method to obtain the content of a certain chapter of the novel

Crawling a certain chapter The content is actually relatively simple:

// 爬取某一章节的内容方法 function getOneChapter(n) { return new Promise((resolve, reject)=>{ if (n >= bookArray.length) { reject('未能找到') } let name = bookArray[n].name; request = https.request(bookArray[n].href, { encoding:'gbk' }, (res)=>{ let html = '' res.on('data', chunk=>{ html += chunk; }) res.on('end', ()=>{ let $ = cheerio.load(html); let content = $("#content").text(); if (content) { // 写成txt writeFun(`./hasGetfile/${name}.txt`, content, 'w') resolve(content); } else { reject('未能找到') } }) }) request.end(); }) } getOneChapter(10)

Example of node crawling data: Let’s talk about how to crawl novel chapters


In this way, you can create a calling interface according to the above method, pass in different chapter parameters, and obtain the current Chapter data

const express = require('express'); const IO = express(); const { getAllChapter, getOneChapter } = require('./readIndex') // 获取章节超链接链表 getAllChapter(); IO.use('/book',function(req, res) { // 参数 let query = req.query; if (query.n) { // 获取某一章节数据 let promise = getOneChapter(parseInt(query.n - 1)); promise.then((d)=>{ res.json({ d: d }) }, (d)=>{ res.json({ d: d }) }) } else { res.json({ d: 404 }) } }) //服务器本地主机的数字 IO.listen('7001',function(){ console.log("启动了。。。"); })

Effect:

Example of node crawling data: Let’s talk about how to crawl novel chapters

Now, a simple chapter search interface is ready, and some parameters can also be judged.

For different data interfaces, the crawler processing methods are also different. However, in the links crawled this time, the display of the content is not dynamically rendered by the front end, so you can directly crawl the static html. Can. If the data is a json string obtained through Ajax or other methods, then the data must be requested through the network interface.

For more node-related knowledge, please visit:

nodejs tutorial!

The above is the detailed content of Example of node crawling data: Let’s talk about how to crawl novel chapters. For more information, please follow other related articles on the PHP Chinese website!

Statement:
This article is reproduced at:juejin.cn. If there is any infringement, please contact admin@php.cn delete