Optimizing Dynamic Pricing for E-commerce

Adomas Sulcas

Apr 21, 2020 8 min read

Pricing intelligence is the process of acquiring and processing data in order to optimize pricing strategies and drive growth in profits. It is quickly becoming a critical part of any consumer-facing  business as dynamic pricing is an increasingly important process in most sectors. While dynamic pricing is just one way to implement strategies developed using pricing intelligence, it is the most common one.

Guide to Dynamic Pricing and Data Acquisition for E-commerce Businesses


This field is required

Thank you!

Please check your inbox.

We will use your email to send you a link to our research material. We will also provide you with information on Oxylabs’ services that may be of interest to you. Be sure that you can opt-out from any marketing related communications that we send you at any time. For more information on your rights and data use please read our Privacy Policy.

Dynamic pricing requires companies to track many external and internal factors related to the valuation of a product or service. Internal factors include:

  • Stock 
  • Production costs 
  • Shipping costs

Meanwhile, external aspects are harder to track as they might vary due to time (seasonal weather changes, holidays, etc.), consumer behavior (such as trends and brand loyalty), competition and many other factors. Yet, most of all, efficient dynamic pricing requires an accurate and effective process of data crawling and parsing.

How dynamic pricing works
Dynamic pricing in a few simple steps

Acquiring accurate data can be challenging due to the constantly changing nature of consumer information. Data extraction can be divided into several different processes that allow for the continuous scanning and scraping of the Internet. These processes are: 

  • Building a scraping path 
  • Creating data extraction scripts
  • Proxy management 
  • Data storage 
  • Data parsing

Each of these steps brings their own challenges, although, implementing a pricing data extraction process isn’t difficult in itself. Issues often arise whenever websites start fighting back against data scraping.

Implementing the pricing data extraction process

In order to apply dynamic pricing, accurate and real-time data is required. Implementing a pricing data extraction process is one of the few ways to continuously acquire such data as only fresh information can be used for dynamic pricing strategies.

Data Acquisition

Pricing data acquisition and dynamic pricing strategies require computing power

Data acquisition is an umbrella term for a 4-step process:

  1. Developing data extraction scripts and implementing them
  2. Using (headless) browsers and their automation
  3. Data parsing
  4. Data storage
Data acquisition is an integral part of dynamic pricing. Steps are developing data extraction scripts, browsers and automation, parsing and storage.
Developing data acquisition tools is a critical part of dynamic pricing strategies

Building a pricing data acquisition tool is generally not too difficult for teams with dedicated and experienced developers. Nowadays at least several languages have easily accessible libraries that make building data extraction scripts and parsing acquired information significantly easier. Additionally, open source tools (e.g. Beautiful Soup) can serve as good starting grounds for some projects.

On the other hand, maintaining a constant flow of pricing data is challenging. Most website owners aren’t too keen to give large swaths of data to anyone with a Python script. As such, most data extraction targets (especially ones that are critical to pricing intelligence) have built algorithms to detect bot-like activity.

Avoiding anti-bot detection algorithms relies on two key factors – understanding how websites detect bot activity and consistently rotating IP addresses through proxies.

Avoiding blocked IPs

Avoiding blocked IPs
Dynamic pricing strategies require continuous access to data sources

There are two ways to consistently scrape the web for pricing intelligence – using residential proxies or using a web scraping solution such as the Real-Time Crawler.

Residential proxies allow web scrapers to acquire IP addresses of physical devices which makes detection significantly harder. These proxies need to be consistently changed in order to avoid bot detection algorithms. Exact details on IP change frequency, work volume etc will depend on the scraping target. 

Efficient web scraping requires not only a large pool of residential proxies in the world but also a large degree of customization:

  • Backconnect entries 
  • City and country-level targeting
  • Session control 
  • Sticky ports.

Managing proxies without a dedicated developer team is a daunting task, therefore many Oxylabs clients choose to outsource their data acquisition pipeline. After all, it is often significantly more time and cost efficient. Our data collection tool is created for those looking to focus mainly on data analysis and developing dynamic pricing strategies.

Going straight to analysis with Real-Time Crawler

Aleksandras Sulzenko explains Real-Time Crawler

Out-of-the-box solutions like Real-Time Crawler bring many benefits to those looking for a reliable way to extract data. Proxy management, data extraction script development and other resource-intensive tasks are done on the side of the provider, allowing companies to focus making use of the data

Additionally, tools like RTC can extract data from challenging targets without the requiring increased resource allocation. Scraping large e-commerce websites or some search engines can require additional development time as these targets often create sophisticated anti-bot algorithms. In-house solutions would have to create workarounds through trial and error which means inevitable slowdowns, blocked IP addresses and an unreliable flow of pricing data.

Tools like Real-Time Crawler make data acquisition easy because:

  • 100% data acquisition success rate
  • Implementing RTC into any project is easy using our API
  • Highly scalable and customizable
  • Receive fully parsed data (in JSON format)
  • Accurate geo-location based pricing data
  • Updates are managed on our side

Oxylabs’ Real-Time Crawler strength lies in its ability to be tailored to almost any business case. Real-Time Crawler can easily scrape the largest e-commerce and search engine targets without any hiccups. It can be easily implemented in place of proxies (through SuperAPI) if a scraping solution has already been built.

Real-Time Crawler in detail

Our Real-Time Crawler is an out-of-the-box solution for businesses looking to cut through the development process and move straight to data analysis. It is well suited for scraping and acquiring data from search engines and large e-commerce websites.

There are two ways to receive parsed data from our RTC tool – Callback or Realtime. Callback allows for both single and batch queries to be sent. Larger queries might take some time and their job status can be checked by sending a GET request. For some tasks (if a callback url is provided) we will send you a message automatically once the task has been completed.

Callback vs. Real-Time: Best Data Delivery Methods
Real-Time Crawler: Callback

On the other hand, the Realtime option is exactly what it says on the tin – we send you the data on an open connection (timeouts are 100 seconds). Ensure that timeout times align as sending data over might take some time.

Real-Time Crawler: Realtime method
Real-Time Crawler: Realtime

By default, all Real-Time Crawler results are stored in our databases which means extra queries. This setting can easily be changed by providing your own cloud storage. You can give our Real-Time Crawler a whirl here!

Real-Time Crawler is an exceptional tool for pricing intelligence as it covers the entire data acquisition and parsing pipeline with 24/7 system monitoring. Resources can be heavily geared towards providing dynamic pricing solutions to clients.

Oxylabs’ Real-Time Crawler saves costs in two ways – by reducing the price of data acquisition and by cutting down on internal spending.

Both residential proxies and Real-Time Crawler work great for data acquisition:

Residential proxiesReal-Time Crawler
FlexibleRequires no infrastructure
More control over developmentPredictable pricing
Can be utilized in other tasks100% success rate
Pay for trafficAccurate data
Advanced configuration controlPay for requests only
Inbuilt JavaScript rendering
No extra scaling costs


Both options have their own benefits and drawbacks, though, many of our clients choose to use our Real-Time Crawler as it allows them to cut down on data extraction costs. Real-Time Crawler makes pricing intelligence almost effortless due to ease-of-use, simple implementation and almost infinite scalability. If you are looking for more efficient ways to extract data from large e-commerce websites and search engines, Real-Time Crawler will be the game changer you need.

Ready to get started? Sign up and test out our solutions by clicking here! Still having some doubts which option is better? Need more technical details before a decision can be made? Contact our sales department by clicking here


About Adomas Sulcas

Adomas Sulcas is a Content Manager at Oxylabs. Having grown up in a tech-minded household, he quickly developed an interest in everything IT and Internet related. When he is not nerding out online or immersed in reading, you will find him on an adventure or coming up with wicked business ideas.

All information on Oxylabs Blog is provided on an "as is" basis and for informational purposes only. We make no representation and disclaim all liability with respect to your use of any information contained on Oxylabs Blog or any third-party websites that may be linked therein. Before engaging in scraping activities of any kind you should consult your legal advisors and carefully read the particular website's terms of service or receive a scraping license.

Related articles

Oxylabs Proxy Integration With Puppeteer

Oxylabs Proxy Integration With Puppeteer

Aug 19, 2021

2 min read