123ArticleOnline Logo
Welcome to 123ArticleOnline.com!
ALL >> Computer-Programming >> View Article

Scrape Amazon Reviews With Scrapy Using Python

Profile Picture
By Author: ReviewGators
Total Articles: 1
Comment this article
Facebook ShareTwitter ShareGoogle+ ShareTwitter Share

We search many things on the internet daily to purchase something, for comparing one product with another, how to decide that a specific product is superior to other? – We straight away check the reviews as well as see how many stars or positive feedbacks have been provided to the products.

Here, we will extract reviews from amazon.com with how many stars it has got, who had posted the reviews, etc.

We would be saving the data in CSV or an excel spreadsheet. The data fields we will scrape include:

Review’s Title
Reviewer’s Name
Review’s Description or Content
Useful Counts
Now, let’s start.

We choose Scrapy – a Python framework used for big-scale web data scraping. Together with it, a few other packages would be needed to extract Amazon products data.

Requests – for sending request of the URL
Pandas – for exporting CSV
Pymy SQL – for connecting My SQL server as well as storing data there
Math – for implementing mathematical operations
You may always install these packages like given below using conda or pip.

pip ...
... install scrapy
conda intall -c conda-forge scrapy
Let’s describe Start URL for scraping sellers’ links

Let’s initially observe what it’s like for scraping Amazon reviews for a single product.

We have taken URL: https://www.amazon.com/dp/B07N9255CG

This will be shown in the below image.

When we go to the reviews section, it’ll appear like an image given below. This might have a few different names within reviews.

However, in case, you carefully review those requests while loading a page as well as play a bit with the previous as well as next page of the reviews, you may have noticed that there’s the post request loading having all the content on a page.

Now, we’ll look at the required payloads and headers for any successful reply. In case, you have properly inspected the pages, then you’ll understand the differences between shifting a page as well as how that reflects on requests passed that.




accept: text/html,*/*

accept-encoding: gzip, deflate, br

accept-language: en-US,en;q=0.9

content-type: application/x-www-form-urlencoded;charset=UTF-8

origin: https://www.amazon.com

referer: https://www.amazon.com/Moto-Alexa-Hands-Free-camera-included/productreviews/B07N9255CG?ie=UTF8&reviewerType=all_reviews

user-agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KHTML, like Gecko)

Chrome/81.0.4044.113 Safari/537.36

x-requested-with: XMLHttpRequest


reviewerType: all_reviews

pageNumber: 2

shouldAppend: undefined

reftag: cm_cr_arp_d_paging_btm_next_2

pageSize: 10

asin: B07N9255CG



accept: text/html,*/*

accept-encoding: gzip, deflate, br

accept-language: en-US,en;q=0.9

content-type: application/x-www-form-urlencoded;charset=UTF-8

origin: https://www.amazon.com

referer: https://www.amazon.com/Moto-Alexa-Hands-Free-camera-included/

ie=UTF8&reviewerType=all_reviews& pageNumber=2

user-agent: Mozilla/5.0 (Windows NT 10.0; Win64; x64)
AppleWebKit/537.36 (KHTML, like Gecko) Chrome/81.0.4044.113 Safari/537.36

x-requested-with: XMLHttpRequest


reviewerType: all_reviews

pageNumber: 2

should spend: undefined

reftag: cm_cr_arp_d_paging_btm_next_2

pageSize: 10

asin: B07N9255CG

The Important Part: CODE or Script
There are a couple of different ways of making the script:

Make a complete Scrapy project
Just make the collection of files to narrow down a project size
During the previous tutorial, we had shown you the whole Scrapy project and information to create and also modify that. We have selected the most pointed way possible this time. We will use a collection of files and all the Amazon reviews will be available there!!

Because we use Scrapy & Python for extracting all the reviews, it’s easier to take the road of the xpath.

The most substantial part of the xpath is catching the pattern. As for copying related xpaths from the Google review window and also paste that, it’s extremely easy but pretty old school as well as also not efficient each time.

That’s what we will do here. We’ll observe the xpath for a similar field, so let’s indicate “Review Title” as well as observe how that makes a pattern or something like that to narrow down the xpath.

There are a couple of examples of similar xpath given here.

Here, you can see that there are parallel attributes to a tag that has data about the “Review Title”.

Therefore, resulting xpath the for Review Titles would be,

Just like that, we’ve provided all the xpaths for different fields that we will scrape.

Review Title: //a[contains(@class,"review-title-content")]/span/text()
Ratings: //a[contains(@title,"out of 5 stars")]/@title
Reviewer’s Name: //div[@id="cm_cr-review_list"]//span[@class="a-profile-name"]/text()
Review Description/Content: //span[contains(@class,"review-text-content")]/span/text()
Helpful Counts: /span[contains(@class,"cr-vote-text")]/text()
Clearly, some joining and stripping to end results with some xpath is important to get ideal data. In addition, don’t overlook to remove additional white spaces.

As we have already seen how to go through the pages as well as also about how to scrape data from them, it is time to collect those!!

Here is the entire code given for scraping all the reviews for a single product!!!

import math, requests, json, pymysql

from scrapy.http import HtmlResponse

import pandas as pd

con = pymysql.connect ( 'localhost', 'root', 'password','database' )

raw_dataframe = [ ]

res = requests.get( 'https://www.amazon.com/Moto-Alexa-Hands-Free-camera-included/ product-reviews/B07N9255CG?ie=UTF8&reviewerType.all_reviews' )

response = HtmlResponse( url=res.url,body=res.content )

product_name = response.xpath( '//h1/a/text()').extract_first( default=' ' ).strip()

total_reviews = response.xpath('//span[contains(text(),"Showing")]/text()').extract_first(default='').strip().split()[-2]]

total_pages = math.ceil(int(total_reviews)/10)

for i in range(0,total_pages):

url = f"https//www.amazon.com/hz/reviews-render/ajax/reviews/get/ref=cm_crarp_d_paging_btm_next_{str(i+2)}"

head = {'accept': 'text/html, */*',

'accept-encoding': 'gzip,deflate,br',

'accept-language': 'en-US,en;q=0.9',

'content-type': 'application/x-www-form-urlencoded;charset=UTF-8', 'origin': 'https://www.amazon.com,


'user-agent': 'Mozilla/5.0(Windows NT 10.0; Win64; x64) AppleWebKit/537.36 (KWH, like Gecko) Chrome/81.0.4044.113 Safari/537.36',

'x- requested-with': 'XMLHttpRequest'


payload = {'reviewerType':'all_reviews'

'pageNumber': i+2,

'shouldAppend': 'undefined',

'reftag': f'cm_crarp_d_paging_btm_next_{str(i+2))',

'pageSize': 10,

'asin': '807N9255C',


res = requests.post(url,headers=head,data=json.dumps(payload))

response = HtmlResponse(url=res.url, body=res.content)

loop = response.xpath('//div[contains(@class,"a-section review")]')

for part in loop:

review_title = part.xpath('.//a[contains(@Class,"review-title-content")]/span/text()').extract_first(default=' ').strip()

rating =part.xpath('.//a[contains(@title,"out of 5 stars")]/@title').extract_first(default=' ').strip().split()[0].strip()

reviewername = part.xpath('.//span[@class."a-profile-name']/text()').extract_first(default=' ').strip()

description =''.join(part. xpath('.//span[contains(@class,"review-text-content")]/span/text()') .extract()).strip()

helpful_count =part.xpath('.//span[contains(@class,"cr-vote-text")]/ text()').extract_first(default ='').strip().split()[0].strip()

raw_dataframe.append([product_name,review_title,rating,reviewer_name, description,helpful_count])

df =pd.Dataframe,(raw_dataframe,columns['Product Name','Review Title','Review Rating','Reviewer Name','Description','Helpful Count' ]),

#inserting into mySQL table


#exporting csv

df.to_csv("amazon reviews.csv",index=None)
Points to Consider While Extracting Amazon Reviews
The entire procedure looks extremely easy to use but there could be some problems while scraping that like response issues, captcha issues, and more. To avoid the same, you need to keep a few VPNs or proxies handy to make the process much smoother.

In addition, there are many times when a website changes its structure. In case, the scraping is going to be a long run, then, you need to always keep the error logs within your script, or error alerts would also help and you could be well aware of it the moment when structure gets changed.

All types of review scraping are much more helpful.

To observe product views by the customers in case, you are the seller on the same website.

Also for monitoring, new party sellers

For creating a dataset that is utilized for research for the academic or business objective?

Contact ReviewGators for more information or ask for a free quote!

More About the Author

We are amongst the leading Review Scraping API Service providers in the world, providing customized review scraping APIs to our clients of all sizes. We utilize the newest technologies dedicated to assisting enterprises in getting well-structured and huge-scale data from the web.

Total Views: 44Word Count: 936See All articles From Author

Add Comment

Computer Programming Articles

1. Ideal Web Development Company In Siliguri
Author: Codepoetry India

2. A Stepwise Guide To Turn On The Auto Hdr Mode On Your Windows 11
Author: robort

3. Build Mobile And Web Apps Using Angularjs Development Services
Author: Rushabh Patel

4. A Complete Guide To Save Locations With Maps On Your Apple Device
Author: robort

5. The Significant Role Of Event-driven Apps In Software Development!
Author: Ash Rakars

6. Here’s How To Use Apple Watch’s Optimized Battery Charging In Watchos 7
Author: robort

7. Home Office Setup Ideas To Boost Your Productivity
Author: robort

8. Nintendo Switch Oled Vs. Others: Which One Should You Buy Right Now?
Author: robort

9. Black Friday 2021 Ad Scans: Start Your Holiday Shopping For Next Week Right Now
Author: Robort

10. Black Friday Apple Watch Deals: Save Big Bucks On Watch Se At Best Buy
Author: robort

11. Best Laptops For Programming By Ignitto
Author: jhon smith

12. The Divi Web Design Company's Advantages
Author: Dylan Alice

13. A Complete Guide To Install And Uninstall The Android Apps On Windows 11
Author: Install and Uninstall the Android Apps

14. How To Extract Data Of Top 10 E-commerce Websites From Indian Market?
Author: iWeb Scraping Services

15. How To Make A Web Scraper With Aws Lambda And The Serverless Framework?
Author: 3i Data Scraping

Login To Account
Login Email:
Forgot Password?
New User?
Sign Up Newsletter
Email Address: