Web Scraping Costco Product Details

Share:

Web Scraping Costco

Costco.com is a simple website; it doesn’t use much JavaScript to display products, making web scraping Costco possible with Python requests.

This tutorial shows you how to scrape Costco product details using Python requests and BeautifulSoup.

Data Scraped from Costco

In this tutorial, you scrape three product details from the Costco search results page:

  • Product Name
  • Price
  • URL

Search results page of Costco

Complete product features are not available on the product results page. Therefore, you need to extract them from the product page. The extracted details depend on the product.

Product page of Costco showing product specifications

Web Scraping Costco: The Environment

You require two external packages for Costco data scraping:

Install the external packages using pip.

You also require two other modules, json and re.

  • The json module allows you to save the extracted details to a JSON file.
  • The re module enables you to use RegEx.

They come with the Python standard library, so you don’t need to install it.

Web Scraping Costco: The Code

Code logic for web scraping Costco

Begin Costco product data scraping by importing the four packages mentioned above.

import requests, re, json


from bs4 import BeautifulSoup

The code also needs to use headers that tell the server that the request originated from a user; this is necessary to bypass anti-scraping measures. Define the headers as a dict.

headers = {
                "accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,"
                "*/*;q=0.8,application/signed-exchange;v=b3;q=0.9",
                "accept-language": "en-GB;q=0.9,en-US;q=0.8,en;q=0.7",
                "dpr": "1",
                "sec-fetch-dest": "document",
                "sec-fetch-mode": "navigate",
                "sec-fetch-site": "none",
                "sec-fetch-user": "?1",
                "upgrade-insecure-requests": "1",
                "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) "
                "AppleWebKit/537.36 (KHTML, like Gecko) Chrome/119.0.0.0 Safari/537.36",
            }

It is better to use functions to keep your code neat; this improves readability and facilitates debugging. Therefore, define four functions:

  • getDetails()
  • saveToJson()
  • getAvailablePages()
  • paginate()

getDetails()

The function accepts a BeautifulSoup object and returns Costco product details.

It is better to check if search results exist before locating product details. To do so, check if there is an ‘h1’ tag with the attribute ‘automation-id’ as ‘noResultsFound.’

if soup.find('h1',{"automation-id":"noResultsFound"}):
        raise Exception("No items found")

If no such tag exists, the function can begin locating the product details:

  1. Locate the product-list
  2. Locate all the products
  3. Extract details of individual products

Finding Product-List

The product list will be inside a ‘div’ element with the class ‘product-list grid.’ Locate it using BeautifulSoup’s find() method.

productGrid = soup.find('div',{'class':'product-list grid'})

Locating All the Products

Use BeautifulSoup’s find_all() method to locate all the products separately, giving you a list.

This list will have the ‘div’ tags containing the product details.

products = productGrid.find_all('div',{'class':'product'})

Extracting Details of Individual Products

Define an empty array that will hold the product details.

productDetails = []

Using the re module, create a RegEx pattern to find the anchor tag containing the product link. This pattern matches the ‘automation-id’ of each product. Using RegEx is necessary as the ‘automation-id’ has a number associated with it that is different for each product.

autoId = re.compile(r"productDescriptionLink_\d")

Iterate through the list, and in each iteration:

  1. Use the RegEx pattern mentioned above to find the required anchor tag.
    link = product.find('a',{'automation-id':autoId})
  2. Extract
    • The product link from the ‘href’ attribute of the anchor tag.
      url = link['href']
    • The name from the tag text.
      name = link.text
    • The price from the text of a ‘div’ tag with the class ‘price.’
      price = product.find('div',{'class':'price'}).text
    • The features from the product page
  3. Append the extracted details to an array.
    productDetails.append(
                   
                    {
                        "Name": name.strip(),
                        "Price": price.strip(),
                        "Features": featuresAsJson,
                        "URL":url
                    }
    

In the second step, extracting features from the product page requires making an HTTP request to the extracted URL.

productResponse = requests.get(url,headers=headers).text

You can then extract the features from the response:

  1. Pass the response text to BeautifulSoup.
    productSoup = BeautifulSoup(productResponse,"lxml")
  2. Locate the ‘div’ containing features; it will have the class ‘product-info-description.’
    featuresDiv = productSoup.find('div',{'class':'product-info-description'})
  3. Locate an unordered list if the above ‘div’ exists; the list will contain the features.
    features = featuresDiv.find('ul').text.split('\n') if featuresDiv.find('ul') else None
  4. Iterate through the list if the features are not None. In each iteration, extract the key-value pairs and store them in a dict featuresAsJson.
    if features != None:
                    for line in features:
                        if line !="":
                            keyNValue = line.split(':')
                            if len(keyNValue) == 2:
                                featuresAsJson[keyNValue[0].strip()] = keyNValue[1].strip()

All products may not have product-info-description. However, all of them have product-info-specs.

Therefore, if the code couldn’t get product-info-description, locate the div tag with the class ‘product-info-specs,’ extract key-value pairs from each row, and store them in the dict.

if featuresAsJson == {}:
                specs = productSoup.find('div',{'class':'product-info-specs'})
                keyNValues = specs.find_all('div',{'class':'row'})
                for keyValue in keyNValues:
                    divs = keyValue.find_all('div')
                    featuresAsJson[divs[0].text.strip()] = divs[1].text.strip()

saveToJson()

This function accepts the extracted product details and saves them to a JSON file. It uses the dump() method of the json module to do so.

with open("costcoProducts.json","w") as f:
        json.dump(productDetails,f,indent=4,ensure_ascii=False)

getAvailablePages()

The code in this tutorial for web scraping Costco can scrape details of all the products from all the pages, but you need to know the number of available pages. This function extracts that number.

The function accepts a query and builds the URL of Costco SERP.

url = f"https://www.costco.com/CatalogSearch?dept=All&keyword={query}"

It then makes an HTTP request to the URL and passes the response text to BeautifulSoup for parsing. This response will contain the homepage’s HTML code, which you can use to find the number of available pages.

response = requests.get(url,headers=headers)
soup = BeautifulSoup(response.text)

The pagination link that takes you to the last page gives the number of available pages. This link comes just before the link that takes you to the next page and has the class ‘forward.’

Therefore, locate the ‘li’ tag with the class ‘forward’ and get its previous sibling.

If the code can’t locate such an element, it initializes available_pages with one.

try:
    availablePages = int(soup.find('li',{'class':'forward'}).previous_sibling.previous_sibling.text.strip()[0])
except:
    availablePages = 1

Note: You can avoid this function, but you need to handle the error whenever the current page number exceeds the number of available pages.  

paginate()

This function accepts three arguments:

  1. allDetails: a dict to save the extracted data from each page.
  2. max_pages: a variable that limits the number of pages scraped.
  3. query: a variable that holds the query obtained from the user input.

The paginate() function starts by initializing ‘page = 1.’ This variable stores the page the code is currently scraping.

Then, the paginate() function runs a while loop as long as ‘page’ is less than ‘max_pages.’

In each loop, the function:

  1. Builds a URL using ‘page’ and ‘query’
    url = f"https://www.costco.com/CatalogSearch?currentPage={page}&dept=All&pageSize=24&keyword={query}"
  2. Increments the page variable by one
    page = page +1
  3. Makes an HTTP request to the URL and creates a BeautiufulSoup object using the response
    response = requests.get(url,headers=headers)
    soup = BeautifulSoup(response.text)
  4. Calls getDetails() with the object as an argument
    productDetails= getDetails(soup)
  5. Saves the extracted details, which will be a list, in the allDetails dict.
    if productDetails !=[]:
        allDetails[f"Page {page-1}"] = productDetails

Now, you can integrate all the functions.

Start by accepting the user’s search query. For queries containing multiple words, split each word and join them using the ‘+’ sign.

product = input("What do you want to buy?")
query = "+".join(product.split())

You can use this query to check for the available pages. Call getAvailablePages() with the query as the argument, which will return the number of available pages. Store it in the variable available_pages.

availablePages = getAvailablePages(query)

Next, accept the maximum number of pages to scrape from the user and store it in the variable mpage.

mpage = int(input("How many pages do you want to scrape?"))

You can now check if mpage is less than or equal to the available_pages.

If so, initialize a variable max_pages with mpage. Otherwise, initialize the variable with available_pages.

max_pages = mpage if mpage <= availablePages else availablePagesnbsp;

Define an empty dict to store the extracted product details from each page. This way, you can store the product details from each page separately; the keys will be the page numbers, and the values will be the product details.

allDetails = {}

You can now call paginate() with allDetails, max_pages, and query as arguments. The function will update the declared allDetails dict.

paginate(allDetails,max_pages,query)

Finally, call the saveToJson() with allDetails as the argument to save extracted data as a JSON file.

saveToJson(allDetails)

The extracted Costco product details will look like this.

"Page 1": [
        {
            "Name": "Kirkland Signature 10-Gallon Wastebasket Liner, Clear, 500-count",
            "Price": "$13.99",
            "Features": {
                "Brand": "Kirkland Signature",
                "Capacity": "10 Gallon",
                "Quantity": "500 Bag(s)"
            },
            "URL": "https://www.costco.com/kirkland-signature-10-gallon-wastebasket-liner%2c-clear%2c-500-count.product.100224137.html"
        },
        {
            "Name": "Kirkland Signature 18-Gallon Compactor & Kitchen Trash Bag, 70-count",
            "Price": "$17.99",
            "Features": {
                "Capacity": "18 Gallon",
                "Thickness": "2.0 mil",
                "Dimensions": "25.625” x 28”"
            },
            "URL": "https://www.costco.com/kirkland-signature-18-gallon-compactor-%2526-kitchen-trash-bag%2c-70-count.product.4000234602.html"
        },
        {
            "Name": "Kirkland Signature Flex-Tech 13-Gallon Kitchen Trash Bag, 200-count",
            "Price": "$19.99",
            "Features": {
                "Brand": "Kirkland Signature",
                "Capacity": "13 Gallon",
                "Quantity": "200 Bag(s)"
            },
            "URL": "https://www.costco.com/kirkland-signature-flex-tech-13-gallon-kitchen-trash-bag%2c-200-count.product.4000101524.html"
        },

And here is the complete code for web scraping Costco product details.

import requests, re, json
from bs4 import BeautifulSoup


headers = {
                "accept": "text/html,application/xhtml+xml,application/xml;q=0.9,image/webp,image/apng,"
                "*/*;q=0.8,application/signed-exchange;v=b3;q=0.9",
                "accept-language": "en-GB;q=0.9,en-US;q=0.8,en;q=0.7",
                "dpr": "1",
                "sec-fetch-dest": "document",
                "sec-fetch-mode": "navigate",
                "sec-fetch-site": "none",
                "sec-fetch-user": "?1",
                "upgrade-insecure-requests": "1",
                "user-agent": "Mozilla/5.0 (Windows NT 10.0; Win64; x64) "
                "AppleWebKit/537.36 (KHTML, like Gecko) Chrome/119.0.0.0 Safari/537.36",
            }


def getDetails(soup):


    if soup.find('h1',{"automation-id":"noResultsFound"}):
        raise Exception("No items found")
   
    productGrid = soup.find('div',{'class':'product-list grid'})
   
    products = productGrid.find_all('div',{'class':'product'})
    productDetails = []
    autoId = re.compile(r"productDescriptionLink_\d")


    for product in products:


        try:
           


            link = product.find('a',{'automation-id':autoId})


            url = link['href']
            print(url)
            name = link.text
           
            price = product.find('div',{'class':'price'}).text


            productResponse = requests.get(url,headers=headers).text
            productSoup = BeautifulSoup(productResponse,"lxml")


            featuresDiv = productSoup.find('div',{'class':'product-info-description'})
            features = featuresDiv.find('ul').text.split('\n') if featuresDiv.find('ul') else None


            featuresAsJson = {}


            if features != None:
                for line in features:
                    if line !="":
                        keyNValue = line.split(':')
                        if len(keyNValue) == 2:
                            featuresAsJson[keyNValue[0].strip()] = keyNValue[1].strip()


            if featuresAsJson == {}:
                specs = productSoup.find('div',{'class':'product-info-specs'})
                keyNValues = specs.find_all('div',{'class':'row'})
                for keyValue in keyNValues:
                    divs = keyValue.find_all('div')
                    featuresAsJson[divs[0].text.strip()] = divs[1].text.strip()
               
            productDetails.append(
               
                {
                    "Name": name.strip(),
                    "Price": price.strip(),
                    "Features": featuresAsJson,
                    "URL":url
                }


            )


        except Exception as e:
            print(e)
            continue
   
    return productDetails


def saveToJson(productDetails):
    with open("costcoProducts.json","w") as f:
        json.dump(productDetails,f,indent=4,ensure_ascii=False)


def getAvailablePages(query):
   
    url = f"https://www.costco.com/CatalogSearch?dept=All&keyword={query}"
    response = requests.get(url,headers=headers)
    soup = BeautifulSoup(response.text)


    try:
        availablePages = int(soup.find('li',{'class':'forward'}).previous_sibling.previous_sibling.text.strip()[0])
    except:
        availablePages = 1


    return availablePages


def paginate(allDetails,max_pages,query):
    page=1
    while page <= max_pages:


        print(page)
        url = f"https://www.costco.com/CatalogSearch?currentPage={page}&dept=All&pageSize=24&keyword={query}"


        print(url)
        page = page +1
        response = requests.get(url,headers=headers)
        soup = BeautifulSoup(response.text)
       
        try:
            productDetails= getDetails(soup)
        except Exception as e:
            print(e)
            continue


        if productDetails !=[]:
            allDetails[f"Page {page-1}"] = productDetails
       


if __name__ == "__main__":


    product = input("What do you want to buy?")
    query = "+".join(product.split())


    availablePages = getAvailablePages(query)


    mpage = int(input("How many pages do you want to scrape?"))
   
    max_pages = mpage if mpage <= availablePages else availablePages


    allDetails = {}
    paginate(allDetails,max_pages,query)


    saveToJson(allDetails)

Code Limitations

The code in this tutorial can scrape product details from Costco SERP. However, keep the following limitations in mind:

  • The code lacks techniques to bypass anti-scraping measures; hence, it is not suitable for large-scale web scraping.
  • Costco may change its website structure, and the code may fail to locate the product details.
  • Some search queries on Costco redirect to specific pages, making it necessary to add additional code.

Wrapping Up

You can scrape Costco product details using Python requests and BeautifulSoup. The process involves making HTTP requests, parsing, and locating the HTML tags holding the required data.

However, the code in this tutorial has some limitations; you need to alter it for large-scale web scraping or when Costco changes its website structure.

If you don’t want to alter the code yourself or want to scrape at a large scale, contact ScrapeHero now.

ScrapeHero is a full-service web scraping service provider. We can build you enterprise-grade web scrapers and crawlers according to your specifications. ScrapeHero services also include monitoring and custom RPA services.

We can help with your data or automation needs

Turn the Internet into meaningful, structured and usable data



Please DO NOT contact us for any help with our Tutorials and Code using this form or by calling us, instead please add a comment to the bottom of the tutorial page for help

Table of content

Scrape any website, any format, no sweat.

ScrapeHero is the real deal for enterprise-grade scraping.

Ready to turn the internet into meaningful and usable data?

Contact us to schedule a brief, introductory call with our experts and learn how we can assist your needs.

Continue Reading

NoSQL vs. SQL databases

Stuck Choosing a Database? Explore NoSQL vs. SQL Databases in Detail

Find out which SQL and NoSQL databases are best suited to store your scraped data.
Scrape JavaScript-Rich Websites

Upgrade Your Web Scraping Skills: Scrape JavaScript-Rich Websites

Learn all about scraping JavaScript-rich websites.
Web scraping with mechanicalsoup

Ditch Multiple Libraries by Web Scraping with MechanicalSoup

Learn how you can replace Python requests and BeautifulSoup with MechanicalSoup.
ScrapeHero Logo

Can we help you get some data?