123ArticleOnline Logo
Welcome to 123ArticleOnline.com!
ALL >> Others >> View Article

How Golang Is Used For Web Scraping With Concurrency?

Profile Picture
By Author: Golang Is Used For web Scraping With Concurrency
Total Articles: 177
Comment this article
Facebook ShareTwitter ShareGoogle+ ShareTwitter Share

Concurrency is required to collect data from several sites and update it. In general, this could create a queue that would have to wait for each page to be completed before moving on to the next. Concurrency will allow you to scrape numerous sites at once.

Data Flow Designs
When scraping many websites, you must wait for each page to finish before moving on to the next. The Data Flow diagram shows how ordinary scraping works.

Data Flow Designs
On other hand, a web scraper may scrape several pages at the same time using parallelism in a single request. The Data flow below explains how this differs from traditional scraping.

Data Flow Designs
Use Cases
There might be various reasons to employ concurrency when scraping. We will try to acquire historical exchange rates for a certain currency in this case. Extracting historical data for a month will require 30 queries, and the loading time will grow in proportion to the number of requests.

Here, we will try to find a historical currency between two currencies with a certain customized data range.

Also, we will try to make ...
... a comparison to equate the speed of the non-currency versus concurrency.

Use Cases
Code Highlights
The final code will be large. Please check the source as there is only an important part mentioned in this blog.

GITHUB Go-currency

To access the page’s content, you can simply utilize the net/http package in Golang and submit the request using the NewRequestWithContext method.

You can use the Go lib goquery to quickly parse the HTML tags using selectors. This source may be used to get more information on how to install it or how to use it.

GITHUB Go-query-master

Code Implementation

Initially, utilize the goquery package to develop a reusable function to fetch the page content and parse it into an HTML document. To utilize this function, simply supply the arguments such as the destination URL, method type, and any other factors that will be required (header, form-data, cookies, etc.)

package utils

import (
"context"
"fmt"
"io"
"net/http"
"net/url"
"strings"
"time"


"github.com/PuerkitoBio/goquery"
)


// GetPage call the client page by HTTP request and extract the body to HTML document.
func GetPage(ctx context.Context, method, siteURL string, cookies []*http.Cookie, headers, formDatas map[string]string, timeout int) (*goquery.Document, []*http.Cookie, error) {
// This function can handle both all methods.
// Initiate this body variable as nil for method that doesn't required body.
body := io.Reader(nil)
// If the request contain form-data, add the form-data parameters to the body.
if len(formDatas) > 0 {
form := url.Values{}
for k, v := range formDatas {
form.Add(k, v)
}


body = strings.NewReader(form.Encode())
}


// Create a new HTTP request with context.
req, err := http.NewRequestWithContext(ctx, method, siteURL, body)
if err != nil {
return nil, nil, fmt.Errorf("failed to create http request context: %w", err)
}
// If the request contain headers, add the header parameters.
if len(headers) > 0 {
for k, v := range headers {
req.Header.Add(k, v)
}
}


// If the request contain cookies, add the cookie parameters.
if len(cookies) > 0 {
for _, c := range cookies {
req.AddCookie(c)
}
}


// Use the default timeout if the timeout parameter isn't configured.
reqTimeout := 10 * time.Second
if timeout != 0 {
reqTimeout = time.Duration(timeout) * time.Second
}


// Use default http Client.
httpClient := &http.Client{
Transport: http.DefaultTransport,
CheckRedirect: nil,
Jar: nil,
Timeout: reqTimeout,
}


// Execute the request.
resp, err := httpClient.Do(req)
if err != nil {
return nil, nil, fmt.Errorf("failed to execute http request: %w", err)
}
// Close the response body
defer func() { _ = resp.Body.Close() }()
// // Parsing response body to HTML document reader.
doc, err := goquery.NewDocumentFromReader(resp.Body)
if err != nil {
return nil, nil, fmt.Errorf("failed to parse html: %w", err)
}


// Return HTML doc, cookies.
return doc, resp.Cookies(), nil

We'll then attempt to parse the currency value. Our service will include criteria such as "from," "to," and "date."

The complete scraping code to obtain the currency based on parameters is shown below. We'll need another method to iterate through the time range arguments because this program only scrapes on a given day.

// getCurrencyHistory gets the currency value on a specific date.
func getCurrencyHistory(ctx context.Context, from, to, date string) (*entities.CurrencyHistory, error) {
urlValues := url.Values{
"from": {to}, // Reverse `from` and `to` due to easily parse the currency value.
"amount": {"1"},
"date": {date},
}


siteURL := fmt.Sprintf("https://www.x-rates.com/historical/?%s", urlValues.Encode())


// Scrape the page.
doc, _, err := utils.GetPage(ctx, http.MethodGet, siteURL, nil, nil, nil, 0)
if err != nil {
return nil, err
}


var currencyHistory *entities.CurrencyHistory


// Scrape the currency value.
doc.Find(".ratesTable tbody tr td").EachWithBreak(func(i int, s *goquery.Selection) bool {
// Scrap the attribute href value from `a` tag HTML.
// https://www.x-rates.com/graph/?from=JPY&to=IDR
// Ignore exists value due to also will check in next line.
href, _ := s.Find("a").Attr("href")


// Reverse `from` and `to` due to easily parse the currency value.
if !strings.Contains(href, "to="+from) {
return true
}


// If the target currency match, scrape the text value.
valueString := s.Find("a").Text()
value, err := strconv.ParseFloat(valueString, 64)
if err != nil {
return true
}


currencyHistory = &entities.CurrencyHistory{
Date: date,
Value: value,
}


return false
})


return currencyHistory, nil
}
The concurrency part must then be implemented. Because you will need to manage the problem, you will iterate the time range argument and utilize error group. The below given is how to final code will look:

/ getCurrencyHistories gets the currencies value on a range date.
func getCurrencyHistories(ctx context.Context, start, end time.Time, from, to string) ([]*entities.CurrencyHistory, error) {
// Get the number of days between start and end date.
days := int(end.Sub(start).Hours()/24) + 1


currencyHistories := make([]*entities.CurrencyHistory, days)


eg, ctx := errgroup.WithContext(ctx)


idx := 0
for d := start; d.After(end) == false; d = d.AddDate(0, 0, 1) {
// Defined new variable to avoid mismatch value when using goroutine.
d := d
i := idx


// Concurrently gets the value on specific date.
eg.Go(func() (err error) {
currencyHistory, err := getCurrencyHistory(ctx, from, to, d.Format("2006-01-02"))
currencyHistories[i] = currencyHistory
return err
})


idx++
}


// Wait all request finished and check the error.
if err := eg.Wait(); err != nil {
return nil, err
}


return currencyHistories, nil
}
The performance of utilizing concurrency vs not using concurrency will be compared. Without concurrency, the code will look like this:

// getCurrencyHistories gets the currencies value on a range date.
func getCurrencyHistories(ctx context.Context, start, end time.Time, from, to string) ([]*entities.CurrencyHistory, error) {
// Get the number of days between start and end date.
days := int(end.Sub(start).Hours()/24) + 1


currencyHistories := make([]*entities.CurrencyHistory, days)


idx := 0
for d := start; d.After(end) == false; d = d.AddDate(0, 0, 1) {


currencyHistory, err := getCurrencyHistory(ctx, from, to, d.Format("2006-01-02"))
if err != nil {
return nil, err
}


currencyHistories[idx] = currencyHistory


idx++
}


return currencyHistories, nil
}
Benchmarking

Multiple requests with varying amounts of date queries (1, 2, 5, 10, 20, and 30) will be tested for non and concurrency. To mimic the benchmark, just launch the service and make a request to a currency history endpoint. The number of inquiries is determined by the number of days that pass between the start and finish dates.

For example, suppose you have ten queries:

v1/currency/history?from=IDR&to=JPY&start_date=2022-03-01&end_date=2022-03-10
Test is performed with the device having the specifications as below:

Mac Mini (M1,200)
Chip Apple M1
16 GB memory
macOS Monterey version 12.3
Internet speed 42.53 (Download) 15.34 (Upload) 29ms (Ping)
Internet region: Indonesia
Internet region
As predicted, the response time for non-concurrency would gradually grow over time as the number of requests increases. However, with concurrency, response time rose somewhat but not significantly.

Note that response times may vary based on the state of the website, traffic, internet speed, geography, and other factors.

For more queries or web scraping services contact iWeb Scraping today!

Request for a quote!

Internet region

More About the Author

iWeb scraping is a leading data scraping company! Offer web data scraping, website data scraping, web data extraction, product scraping and data mining in the USA, Spain.

Total Views: 175Word Count: 1479See All articles From Author

Add Comment

Others Articles

1. Boost Your Farm Efficiency With Vishal 328 Leal
Author: Ajay tiwari

2. Effortless Office And Budget Relocations In Auckland
Author: Accurate Movers

3. Stainless Steel Shower Tray Auckland: Strong, Sleek Solutions
Author: Kiwi Stainless

4. Best Cash For Cars In Auckland
Author: Cars 4 Cash

5. Trusted Junk Car Buyers In Auckland
Author: Cars 4 Cash

6. Cryptocurrency In Medical Billing: A Revolutionary Approach
Author: Albert brown

7. Stainless Steel Auckland Kitchen Makeover
Author: Kiwi Stainless

8. The Long-term Effects Of Childhood Trauma And How To Heal
Author: ImPerfect

9. Top It Companies In India: Driving Global Innovation
Author: Johnny

10. Uk Health And Wellness: Trends, Challenges, And Opportunities
Author: Alex

11. Trending Job Categories On Aamdhane
Author: Aamdhane

12. Enhance Your Hearing Experience With Custom Made Ear Plugs And Hearing Aid Solutions In Cardiff
Author: Steve Jones

13. Industrial Hose Pipe Couplings - Manxpower
Author: MANXPOWER

14. Best Astrologer In Koppal
Author: Astrology56

15. Unveiling Photo Editing Tricks: What You Need To Know
Author: Sam

Login To Account
Login Email:
Password:
Forgot Password?
New User?
Sign Up Newsletter
Email Address: