How to Scrape Store Locations from using Python and LXML

Web scraping is a faster and efficient way to get details of store locations for a particular website rather than taking the time to gather details manually. This tutorial is about scraping store locations and contact details available on, one of the largest discount store retailers in the U.S.

For this tutorial, our scraper will extract the details of store information by a given zip code. 

Here is a list of fields we will be extracting:

  1. Store Name
  2. Address
  3. Week Day
  4. Hours Open
  5. Phone Number

Below is a screenshot of the data that will be extracted as part of this tutorial.

There is a lot more data we could scrape from the store detail page on Target such as pharmacy and grocery timings, but for now, we’ll stick to these.

Scraping Logic

  1. Construct the URL of the search results page from Target. Let’s take the location Clinton, New York. We’ll have to create this URL manually to scrape results from that page-
  2. Download HTML of the search result page using Python Requests – Quite easy, once you have the URL. We use Python requests to download the entire HTML of this page.
  3. Parse the page using LXML – LXML lets you navigate the HTML Tree Structure using Xpaths. We have predefined the XPaths for the details we need in the code.
  4. Save the data to a JSON file.


  • Python 2.7 ( )
  • PIP to install the  following packages in Python (
  • Python Requests, to make requests and download the HTML content of the pages (
  • Python LXML, for parsing the HTML Tree Structure using Xpaths ( Learn how to install that here – )
  • UnicodeCSV for handling Unicode characters in the output file. Install it using pip install unicodecsv.

The Code


If the embed above does not work, you can click on the link here.

Running the Scraper

Assuming the scraper is named If you type in the script name in command prompt along with a -h

usage: [-h] zipcode

positional arguments:
zipcode Zip code

optional arguments:
-h, --help show this help message and exit

The argument zip code is the zip code to find the stores near the particular location.

As an example, to find all the target stores in and near Clinton, New York we would put the argument as 12901 for zip code:

python 12901

This will create a JSON output file called 12901-locations.json that will be in the same folder as the script.

The output file will look similar to this.

        "County": "Clinton", 
        "Store_Name": "Plattsburgh", 
        "State": "NY", 
        "Street": "60 Smithfield Blvd", 
        "Stores_Open": [
        "Contact": "(518) 247-4961", 
        "City": "Plattsburgh", 
        "Country": "United States", 
        "Zipcode": "12901-2151", 
        "Timings": [
                "Week Day": "Monday-Friday", 
                "Open Hours": "8:00 a.m.-10:00 p.m."
                "Week Day": "Saturday", 
                "Open Hours": "8:00 a.m.-10:00 p.m."
                "Week Day": "Sunday", 
                "Open Hours": "8:00 a.m.-9:00 p.m."

You can download the code at

Let us know in the comments below how this scraper worked for you.

Known Limitations

This code should work for extracting details of Target stores for all zip codes available on  If you want to scrape the details of thousands of pages you should read  Scalable do-it-yourself scraping – How to build and run scrapers on a large scale and How to prevent getting blacklisted while scraping.

If you need professional help with scraping complex websites, contact us by filling up the form below.

Tell us about your complex web scraping projects

Disclaimer: Any code provided in our tutorials is for illustration and learning purposes only. We are not responsible for how it is used and assume no liability for any detrimental usage of the source code. The mere presence of this code on our site does not imply that we encourage scraping or scrape the websites referenced in the code and accompanying tutorial. The tutorials only help illustrate the technique of programming web scrapers for popular internet websites. We are not obligated to provide any support for the code, however, if you add your questions in the comments section, we may periodically address them.


Join the conversation