Proxy locations

Europe

North America

South America

Asia

Africa

Oceania

See all locations

Network statusCareers

Back to blog

Scrapy Playwright Tutorial: How to Scrape JavaScript Websites

Scrapy Playwright Tutorial: How to Scrape JavaScript Websites

Roberta Aukstikalnyte

2023-09-284 min read
Share

Scrapy is a popular Python package commonly used in the web scraping field. While it’s ideal for scraping static websites, the same could not be said for dynamic, JavaScript-heavy websites: it may require rendering and additional user input. 

As a result, Scrapy Playwright was released – an advanced headless browser designed for scraping dynamic JavaScript websites. In today’s article, we’ll demonstrate how to successfully gather public data using Playwright integration for Scrapy paired Oxylabs’ proxies.

Let’s get started! 

1. Installing Scrapy Playwright

First, let’s install the Scrapy Playwright pip package:

pip install scrapy-playwright

Now, we can install all the prerequisite browser engines:

playwright install

2. Setting up Scrapy Playwright

Now that we have the Playwright integration for Scrapy installed, let’s set up a basic Scrapy spider using the following command: 

scrapy startproject scrapy_book_crawler
cd scrapy_book_crawler
scrapy genspider books https://books.toscrape.com/js/

This command will initialize a base Scrapy project, navigate to it and create a spider named books for us to work with.

3. Reviewing main settings

To finish setting up our Playwright integration, we need to add some specific Scrapy settings in our settings.py file.

DOWNLOAD_HANDLERS = {
    "http": "scrapy_playwright.handler.ScrapyPlaywrightDownloadHandler",
    "https": "scrapy_playwright.handler.ScrapyPlaywrightDownloadHandler",
}

PLAYWRIGHT_DEFAULT_NAVIGATION_TIMEOUT = (
    30 * 1000
)

PLAYWRIGHT_BROWSER_TYPE = "chromium"
  • In DOWNLOAD_HANDLERS, we specify that we’ll want to use the Scrapy Playwright request handlers for both our http and https requests.

  • In PLAYWRIGHT_DEFAULT_NAVIGATION_TIMEOUT, we can specify the timeout value to be used when requesting pages with Playwright. The default value for Playwright timeouts is 30 seconds. 

  • In PLAYWRIGHT_BROWSER_TYPE, we define the type of browser engine to be used to process requests.

For more in-depth information about these and all the other possible configuration parameters, you can check out the Playwright documentation on GitHub.

4. Scraping with Scrapy Playwright

Let’s start with a simple scraping task just to see if our setup is correct. We’ll gather book titles and prices from https://books.toscrape.com/, a mock website designed for people to practice web scraping. 

First, we’ll create a model for the information that we want scraped:

import scrapy

class Book(scrapy.Item):
   title = scrapy.Field()
   price = scrapy.Field()

We should put this model in the file items.py, which was already created by the Scrapy project initializer. 

Here, we’ve created an Item in the context of Scrapy and defined some fields for it. An Item basically defines the data that we’ll want to extract during our scraping.

Next, let’s add some logic to our spider in spiders/books.py.

import scrapy

from scrapy_book_crawler.items import Book


class BooksSpider(scrapy.Spider):
   """Class for scraping books from https://books.toscrape.com/"""

   name = "books"

   def start_requests(self):
       url = "https://books.toscrape.com/"
       yield scrapy.Request(
           url,
           meta=dict(
               playwright=True,
               playwright_include_page=True,
               errback=self.errback,
           ),
       )

   async def parse(self, response):
       page = response.meta["playwright_page"]
       await page.close()

       for book in response.css("article.product_pod"):
           book = Book(
               title=book.css("h3 a::attr(title)").get(),
               price=book.css("p.price_color::text").get(),
           )
           yield book

   async def errback(self, failure):
       page = failure.request.meta["playwright_page"]
       await page.close()

In the start_requests(self) function, we define the URL and the meta parameters for our requests. To use Playwright, we need to add some variables to our meta dictionary:

  • playwright=True indicates that Scrapy should use Playwright to process this request.

  • playwright_include_page=True makes sure that we can access the Page Playwright object when processing the request and add PageMethods, which will be used later on.

  • errback=self.errback function would close the pages for us in case the request fails at some point.

  • In the parse(self, response) function, we instruct the Page to close after processing, extract the data for the Book item, and yield it.

Now, we can run our spider with the following command and inspect the results:

scrapy crawl books -o books.json

This will save all of the items yielded by Scrapy into a json file.

scrapy

5. Scraping multiple pages 

To gather data from several pages, we need Scrapy to yield an additional request for each URL we find inside our current one.

Let’s modify our parse function to do just that.

async def parse(self, response):
       page = response.meta["playwright_page"]
       await page.close()

       for book in response.css("article.product_pod"):
           book = Book(
               title=book.css("h3 a::attr(title)").get(),
               price=book.css("p.price_color::text").get(),
           )
           yield book
          
       next_page = response.css("li.next a::attr(href)").get()
       if next_page:
           next_page_url = response.urljoin(next_page)
           yield scrapy.Request(
               next_page_url,
               meta=dict(
                   playwright=True,
                   playwright_include_page=True,
                   errback=self.errback,
               ),
           )

Our newly added code will now find the next page URL within our currently processed one. Also, it will yield an additional request for the page to our Scrapy pipeline to be processed. This process will repeat until there are no pages left.

If we run the code with the same command:

scrapy crawl books -o books.json

We can see that there are many more books in our result file now, as our spider has gone through multiple pages to fetch them.

book list

6. Interacting with page

Until now, we haven’t made the most of the potential that Playwright can give. Let’s do something like clicking a category link and taking a screenshot of the rendered webpage.

If we open the page we are scraping, we can see that there’s a list of categories on the right, and each category is clickable.

bookstoscrape

Let’s create a Python enum for all the categories so that we can access them in the code conveniently later on. To do that, we’ll create a file named enums.py and add the enum there:

from enum import Enum


class BookCategory(int, Enum):
   """All category li nth:child numbers for books.toscrape.com"""

   TRAVEL = 1
   MYSTERY = 2
   HISTORICAL_FICTION = 3
   SEQUENTIAL_ART = 4
   CLASSICS = 5
   PHILOSOPHY = 6
   ROMANCE = 7
   WOMENS_FICTION = 8
   FICTION = 9
   CHILDRENS = 10
   RELIGION = 11
   NONFICTION = 12
   MUSIC = 13
   DEFAULT = 14
   SCIENCE_FICTION = 15
   SPORTS_AND_GAMES = 16
   ADD_A_COMMENT = 17
   FANTASY = 18
   NEW_ADULT = 19
   YOUNG_ADULT = 20
   SCIENCE = 21
   POETRY = 22
   PARANORMAL = 23
   ART = 24
   PSYCHOLOGY = 25
   AUTOBIOGRAPHY = 26
   PARENTING = 27
   ADULT_FICTION = 28
   HUMOR = 29
   HORROR = 30
   HISTORY = 31
   FOOD_AND_DRINK = 32
   CHRISTIAN_FICTION = 33
   BUSINESS = 34
   BIOGRAPHY = 35
   THRILLER = 36
   CONTEMPORARY = 37
   SPIRITUALITY = 38
   ACADEMIC = 39
   SELF_HELP = 40
   HISTORICAL = 41
   CHRISTIAN = 42
   SUSPENSE = 43
   SHORT_STORIES = 44
   NOVELS = 45
   HEALTH = 46
   POLITICS = 47
   CULTURAL = 48
   EROTICA = 49
   CRIME = 50

Now, we’ll adjust our start_requests(self) function to make use of some PageMethods.

def start_requests(self):
       category_number = BookCategory.CLASSICS
       url = "https://books.toscrape.com/"
       yield scrapy.Request(
           url,
           meta=dict(
               playwright=True,
               playwright_include_page=True,
               playwright_page_methods=[
                   PageMethod(
                       "click",
                       selector=f"div.side_categories > ul > li > ul > li:nth-child({self.category_number.value}) > a",
                   ),
                   PageMethod("wait_for_selector", "article.product_pod"),
               ],
               errback=self.errback,
           ),
       )

PageMethod allows us to interact with the page before we finish rendering it and sending it as our response for parsing. We specify one method to click the category we have selected from our enum in the variable category_number and another method to wait for a CSS selector of a product listing to load.

Next, we’ll update the parse(self, response) function to take a screenshot of the rendered page using the page.screenshot function and save it as a file:

   async def parse(self, response):
      category_number = BookCategory.CLASSICS
   page = response.meta["playwright_page"]
       await page.screenshot(path=f"books-{self.category_number.name}.png")
       await page.close()

       for book in response.css("article.product_pod"):
           book = Book(
               title=book.css("h3 a::attr(title)").get(),
               price=book.css("p.price_color::text").get(),
           )
           yield book

The full code for our Scrapy spider should look like this:

import scrapy
from scrapy_playwright.page import PageMethod
from scrapy_book_crawler.items import Book
from scrapy_book_crawler.enums import BookCategory


class BooksSpider(scrapy.Spider):
   """Class for scraping books from https://books.toscrape.com/"""

   name = "books"
   category_number = BookCategory.CLASSICS

   def start_requests(self):
       url = "https://books.toscrape.com/"
       yield scrapy.Request(
           url,
           meta=dict(
               playwright=True,
               playwright_include_page=True,
               playwright_page_methods=[
                   PageMethod(
                       "click",
                       selector=f"div.side_categories > ul > li > ul > li:nth-child({self.category_number.value}) > a",
                   ),
                   PageMethod("wait_for_selector", "article.product_pod"),
               ],
               errback=self.errback,
           ),
       )

   async def parse(self, response):
       page = response.meta["playwright_page"]
       await page.screenshot(path=f"books-{self.category_number.name}.png")
       await page.close()

       for book in response.css("article.product_pod"):
           book = Book(
               title=book.css("h3 a::attr(title)").get(),
               price=book.css("p.price_color::text").get(),
           )
           yield book

   async def errback(self, failure):
       page = failure.request.meta["playwright_page"]
       await page.close()

If we run it with this command:

scrapy crawl books -o books.json

We’ll see our screenshot saved within the project files and our results file populated with books from the selected category:

classics

7. Using proxies with Scrapy Playwright

Scrapy Playwright works best with proxies – they can enhance your anonymity by concealing your real IP address and location, thus increasing your chances for successful and block-free scraping operations. 

For the sake of this tutorial, we’re going to use Oxylabs’ Dedicated Datacenter Proxies, but feel free to use another type if you want.  

You can set up proxies for your Playwright spiders in three different ways: 

1) Spider class custom_settings

We can add the proxy settings as launch options within the custom_settings parameter used by the Scrapy Spider class:

class BooksSpider(scrapy.Spider):
   """Class for scraping books from https://books.toscrape.com/"""

   name = "books"
   custom_settings = {
       "PLAYWRIGHT_LAUNCH_OPTIONS": {
           "proxy": {
               "server": "127.0.0.1:60000",
               "username": "username",
               "password": "password",
           },
       }
   }

2) Meta dictionary in start_requests

We can also define the proxy within the start_requests function by passing it within the meta dictionary:

def start_requests(self) -> Generator[scrapy.Request, None, None]:
       url = "https://books.toscrape.com/"
       yield scrapy.Request(
           url,
           meta=dict(
               playwright=True,
               playwright_include_page=True,
               playwright_context_kwargs={
                   "proxy": {
                       "server": "127.0.0.1:60000",
                       "username": "username",
                       "password": "password",
                   },
               },
               errback=self.errback,
           ),
       )

3) PLAYWRIGHT_CONTEXTS in settings.py 

Lastly, we can also define the proxy we want to use within the Scrapy settings file:

PLAYWRIGHT_CONTEXTS = {
   "default": {
       "proxy": {
           "server": "127.0.0.1:60000",
           "username": "username",
           "password": "password",
       },
   },
   "alternative": {
       "proxy": {
           "server": "127.0.0.1:60001",
           "username": "username",
           "password": "password",
       },
   },

And there you have it – we’ve successfully built a scraper that utilizes Playwright to render and interact with JavaScript-enabled websites.

Frequently asked questions

Is Playwright good for scraping?

Yes, Playwright is an excellent choice for scraping, especially when it comes to JavaScript-heavy, dynamic website data. Playwright offers a powerful automation framework that allows you to interact with web pages, click buttons, fill forms, and easily extract data from JavaScript-rendered content. 

Is Playwright better than Selenium?

Both Playwright and Selenium have their advantages. However, when it comes to web scraping, Playwright is often considered a better choice – it offers a user-friendly API with improved speed and better support for multiple browser engines, including Chromium, Firefox, and WebKit. Also, Playwright offers various automation capabilities, e.g., interacting with JavaScript websites. 

Can Scrapy parse JavaScript?

Scrapy cannot render JavaScript on its own – it has to rely on integrations like Playwright and Splash to accomplish this.

What is the difference between Playwright and Scrapy Splash?

Playwright supports a full browser engine like Chrome or Firefox, while Splash uses a webkit-embedded browser.

About the author

Roberta Aukstikalnyte

Senior Content Manager

Roberta Aukstikalnyte is a Senior Content Manager at Oxylabs. Having worked various jobs in the tech industry, she especially enjoys finding ways to express complex ideas in simple ways through content. In her free time, Roberta unwinds by reading Ottessa Moshfegh's novels, going to boxing classes, and playing around with makeup.

All information on Oxylabs Blog is provided on an "as is" basis and for informational purposes only. We make no representation and disclaim all liability with respect to your use of any information contained on Oxylabs Blog or any third-party websites that may be linked therein. Before engaging in scraping activities of any kind you should consult your legal advisors and carefully read the particular website's terms of service or receive a scraping license.

Related articles

Get the latest news from data gathering world

I’m interested

Scale up your business with Oxylabs®