Web Scraping a Go
First steps
First of all we must have Go installed, Instructions to download and install Go.
We create a new folder for the project, move to the directory and execute the following command:
go mod init scraper
? The go mod init command is used to initialize a new Go module in the directory where it is run and creates a go.mod file to track code dependencies. Dependency management
Now let's install Colibri:
go get github.com/gonzxlez/colibri
? Colibri is a Go package that allows us to crawl and extract structured data on the web using a set of rules defined in JSON. Repository
Extraction rules
We define the rules that colibri will use to extract the data we need. Documentation
We are going to make an HTTP request to the URL https://pkg.go.dev/search?q=xpath which contains the results of a query for Go packages related to xpath in Go Packages.
Using the development tools included in our web browser, we can inspect the HTML structure of the page. What are the browser development tools?
<div class="SearchSnippet"> <div class="SearchSnippet-headerContainer"> <h2> <a href="/github.com/antchfx/xpath" data-gtmc="search result" data-gtmv="0" data-test-id="snippet-title"> xpath <span class="SearchSnippet-header-path">(github.com/antchfx/xpath)</span> </a> </h2> </div> <div class="SearchSnippet-infoLabel"> <a href="/github.com/antchfx/xpath?tab=importedby" aria-label="Go to Imported By"> <span class="go-textSubtle">Imported by </span><strong>143</strong> </a> <span class="go-textSubtle">|</span> <span class="go-textSubtle"> <strong>v1.2.5</strong> published on <span data-test-id="snippet-published"><strong>Oct 26, 2023</strong></span> </span> <span class="go-textSubtle">|</span> <span data-test-id="snippet-license"> <a href="/github.com/antchfx/xpath?tab=licenses" aria-label="Go to Licenses"> MIT </a> </span> </div> </div>
Fragment of the HTML structure that represents a result of the query.
Then we need a selector “packages” that will find all the div elements in the HTML with the class SearchSnippet, from those elements a selector “name” will take the text of the element a inside an element h2 and a selector “path” will take the value of the href attribute of the a element within an h2 element. . In other words, “name” will take the name of the Go package and “path” the path of the package :)
{ "method": "GET", "url": "https://pkg.go.dev/search?q=xpath", "timeout": 10000, "selectors": { "packages": { "expr": "div.SearchSnippet", "all": true, "type": "css", "selectors": { "name": "//h2/a/text()", "path": "//h2/a/@href" } } } }
- method: specifies the HTTP method (GET, POST, PUT, ...).
- url: URL of the request.
- timeout: timeout in milliseconds for the HTTP request.
-
selectors: selectors.
-
“packages”: is the name of the selector.
- expr: selector expression.
- all: specifies that all elements matching the expression should be found.
- type: the type of the expression, in this case a CSS selector.
-
selectors: nested selectors.
- “name” and “path” are the names of the selectors and their values are expressions, in this case XPath expressions.
-
“packages”: is the name of the selector.
Code in Go
We are ready to create a scraper.go file, import the necessary packages and define the main function:
package main import ( "encoding/json" "fmt" "github.com/gonzxlez/colibri" "github.com/gonzxlez/colibri/webextractor" ) var rawRules = `{ "method": "GET", "url": "https://pkg.go.dev/search?q=xpath", "timeout": 10000, "selectors": { "packages": { "expr": "div.SearchSnippet", "all": true, "type": "css", "selectors": { "name": "//h2/a/text()", "path": "//h2/a/@href" } } } }` func main() { we, err := webextractor.New() if err != nil { panic(err) } var rules colibri.Rules err = json.Unmarshal([]byte(rawRules), &rules) if err != nil { panic(err) } output, err := we.Extract(&rules) if err != nil { panic(err) } fmt.Println("URL:", output.Response.URL()) fmt.Println("Status code:", output.Response.StatusCode()) fmt.Println("Content-Type", output.Response.Header().Get("Content-Type")) fmt.Println("Data:", output.Data) }
? WebExtractor are default interfaces for Colibri ready to start crawling or extracting data on the web.
Using the New function of webextractor, we generate a Colibri structure with what is necessary to start extracting data.
Then we convert our rules in JSON to a Rules structure and call the Extract method sending the rules as arguments.
We obtain the output and the URL of the HTTP response, the HTTP status code, the content type of the response and the data extracted with the selectors are printed on the screen. See the documentation for the Output structure.
We execute the following command:
go mod tidy
? The go mod tidy command makes sure that the dependencies in the go.mod match the module source code.
Finally we compile and run our code in Go with the command:
go run scraper.go
Conclusion
In this post, we have learned how to perform Web Scraping in Go using the Colibri package, defining extraction rules with CSS and XPath selectors. Colibri emerges as a tool for those looking to automate web data collection in Go. Its rules-based approach and ease of use make it an attractive option for developers of all experience levels.
In short, Web Scraping in Go is a powerful and versatile technique that can be used to extract information from a wide range of websites. It is important to highlight that Web Scraping must be carried out ethically, respecting the terms and conditions of the websites and avoiding overloading their servers.
The above is the detailed content of Web Scraping a Go. For more information, please follow other related articles on the PHP Chinese website!

Hot AI Tools

Undresser.AI Undress
AI-powered app for creating realistic nude photos

AI Clothes Remover
Online AI tool for removing clothes from photos.

Undress AI Tool
Undress images for free

Clothoff.io
AI clothes remover

AI Hentai Generator
Generate AI Hentai for free.

Hot Article

Hot Tools

Notepad++7.3.1
Easy-to-use and free code editor

SublimeText3 Chinese version
Chinese version, very easy to use

Zend Studio 13.0.1
Powerful PHP integrated development environment

Dreamweaver CS6
Visual web development tools

SublimeText3 Mac version
God-level code editing software (SublimeText3)

Hot Topics



OpenSSL, as an open source library widely used in secure communications, provides encryption algorithms, keys and certificate management functions. However, there are some known security vulnerabilities in its historical version, some of which are extremely harmful. This article will focus on common vulnerabilities and response measures for OpenSSL in Debian systems. DebianOpenSSL known vulnerabilities: OpenSSL has experienced several serious vulnerabilities, such as: Heart Bleeding Vulnerability (CVE-2014-0160): This vulnerability affects OpenSSL 1.0.1 to 1.0.1f and 1.0.2 to 1.0.2 beta versions. An attacker can use this vulnerability to unauthorized read sensitive information on the server, including encryption keys, etc.

The article explains how to use the pprof tool for analyzing Go performance, including enabling profiling, collecting data, and identifying common bottlenecks like CPU and memory issues.Character count: 159

The article discusses writing unit tests in Go, covering best practices, mocking techniques, and tools for efficient test management.

Queue threading problem in Go crawler Colly explores the problem of using the Colly crawler library in Go language, developers often encounter problems with threads and request queues. �...

The library used for floating-point number operation in Go language introduces how to ensure the accuracy is...

Backend learning path: The exploration journey from front-end to back-end As a back-end beginner who transforms from front-end development, you already have the foundation of nodejs,...

The article discusses the go fmt command in Go programming, which formats code to adhere to official style guidelines. It highlights the importance of go fmt for maintaining code consistency, readability, and reducing style debates. Best practices fo

Under the BeegoORM framework, how to specify the database associated with the model? Many Beego projects require multiple databases to be operated simultaneously. When using Beego...
