Give us a call: (800) 252-6164
Hedge Funds · Alternative Data · Demand Measurement

DEMAND SIGNALS
What to Crawl to Measure Real-Time Demand

Reported revenue is lagging. Demand is revealed continuously across the web. Potent Pages builds durable crawling and extraction systems that convert pricing, inventory, rankings, and engagement behavior into structured time-series demand signals your fund can backtest, monitor, and trade around.

  • Detect inflections earlier
  • Measure demand at SKU / geo level
  • Separate demand vs supply constraints
  • Deliver clean time-series outputs

Why real-time demand matters

Markets are forward-looking, but most demand data is backward-looking. Revenue is reported weeks after the quarter ends. Surveys lag, vendor dashboards smooth, and consensus forms quickly. The research edge moves upstream: the goal is to observe demand as it forms, shifts, and accelerates—before it is captured in financial statements.

Key idea: Demand leaves a continuous footprint on the public web—search behavior, inventory depletion, pricing response, and engagement all update in near real time.

What “demand signals” mean in web data

A demand signal is an observable web-based behavior that correlates with underlying economic demand. The most useful signals measure change over time: velocity, acceleration, and dispersion across geographies, channels, or product tiers.

Discovery & intent

Search rankings, query modifiers, category navigation changes, and “compare” behavior that precede transactions.

Scarcity & fulfillment

Stock-outs, backorders, delivery-date shifts, and regional availability that reveal demand pressure.

Pricing response

Discount depth and cadence, dynamic pricing, bundling, and promo intensity as real-time demand feedback loops.

Engagement & conversion friction

Waitlists, “notify me,” checkout changes, and queue systems that separate latent demand from operational constraints.

Practical insight: In most categories, you get better signal by crawling a smaller, thesis-driven universe at higher frequency than by crawling everything at low cadence.

A practical demand-crawling framework

Real-time demand measurement works best when it is approached like research infrastructure—not one-off scraping. A strong framework starts with a hypothesis, then maps it to a measurable proxy and collection strategy that preserves comparability.

1

Define the demand question

What is changing—category demand, brand preference, willingness to pay, or conversion readiness?

2

Choose the observable proxy

Inventory depletion, delivery-date drift, price response, ranking movement, review velocity, or funnel friction.

3

Select sources & cadence

Prioritize high-signal pages and crawl frequently enough to capture inflections, not snapshots.

4

Normalize into a stable schema

Unify messy sources into time-series tables with versioned definitions and consistent units.

5

Engineer investable indicators

Transform raw changes into velocity, acceleration, dispersion, and anomaly metrics aligned to your horizon.

6

Monitor, repair, and iterate

Preserve continuity through site changes and refine definitions as you learn where signal-to-noise improves.

What to crawl: the highest-signal demand sources

Below are the most common web-based demand sources used by hedge funds. The goal is to capture change: what moved, how fast, and where.

Search & discovery surfaces

Rankings by keyword and geography, category navigation shifts, and emerging query modifiers that indicate demand formation.

Product pages (brand + retailer)

In-stock status, “notify me” prompts, delivery estimates, variant availability, and SKU lifecycle changes.

Pricing & promotions

Price moves, markdown depth, promo cadence, bundling, and dynamic pricing behavior—captured at SKU and channel level.

Marketplaces & aggregators

Bestseller ranks, category movement, seller churn, fulfillment timing, and assortment expansion or contraction.

Reviews, Q&A, and support

Review volume velocity, question frequency, complaint clustering, and substitution language that signals demand shift.

Enterprise demand proxies

Pricing page changes, documentation updates, demo/CTA patterns, and hiring velocity tied to sales and implementation.

Implementation note: Many teams start with daily crawling, then increase cadence around catalysts (earnings, product launches, promotions, supply disruptions).

Separating demand from supply constraints

One of the most valuable outcomes of demand crawling is distinguishing true demand shifts from supply-side noise. Stock-outs, delays, and promo intensity can look like demand signals unless you capture the surrounding context.

  • Demand pressure: stock-outs + rising prices + stable engagement often indicate scarcity-driven strength.
  • Weak demand: persistent discounting + stable inventory + declining engagement often indicate softness.
  • Operational constraints: waitlists/queues + stable pricing may indicate capped demand rather than weak demand.
  • Channel shift: brand site weakness alongside marketplace strength may indicate distribution reallocation.
Key idea: Triangulation across sources (brand + retailer + marketplace) is how you reduce false positives.

What makes a demand signal investable

The best demand indicators are not just predictive in a backtest—they are operationally stable in production. Investability comes from a combination of economics, data integrity, and disciplined definitions.

Low latency

Updates frequently enough to matter for your horizon (intraday, daily, weekly), especially around catalysts.

Continuity

Long-running collection with monitoring and repair workflows that preserve time-series comparability.

Stable schemas

Versioned definitions so research and production outputs do not silently drift.

Backtest-ready output

Structured tables, anomaly flags, and clear lineage—not raw HTML dumps.

Warning: Many demand signals “work” once. Durable signals survive website changes and shifting product catalogs.

Questions About Demand Signals & Real-Time Demand Data

These are common questions hedge funds ask when exploring demand proxies, web crawling, and alternative data pipelines built for backtesting and production monitoring.

What is a “demand signal” in alternative data? +

A demand signal is an observable, repeatable web-based behavior that correlates with underlying economic demand. The highest-value signals measure change over time—for example, inventory depletion velocity, delivery-date drift, discount cadence, marketplace rank movement, or review volume acceleration.

Rule of thumb: signals that capture velocity and inflection are usually more tradable than one-off snapshots.
What should we crawl to measure consumer demand in real time? +

Most consumer demand stacks combine three layers: discovery, transaction-adjacent behavior, and post-purchase activity.

  • Discovery: search and category navigation movement
  • Transaction-adjacent: product pages, inventory status, delivery estimates
  • Post-purchase: review velocity, Q&A volume, support complaints

The strongest setups crawl across multiple channels (brand, retailers, marketplaces) to separate demand from distribution noise.

How do you separate demand strength from supply constraints? +

You separate them by capturing context across signals. For example, a stock-out with rising prices and stable engagement looks different than a stock-out with aggressive discounting or declining traffic.

  • Pair inventory signals with pricing response
  • Compare brand site vs retailer vs marketplace behavior
  • Track delivery-date drift and “notify me” prompts as demand-pressure indicators
Why build bespoke crawlers instead of buying a vendor demand dataset? +

Vendor datasets optimize for resale and broad coverage, which often means lower frequency, opaque methodology, and diluted edge. Bespoke crawling lets your fund define:

  • Universe and SKU basket selection
  • Cadence (including event-driven crawling)
  • Normalization rules and schema versioning
  • Signal features aligned to your horizon
Practical benefit: control reduces vendor opacity risk and preserves continuity across site changes.
What does Potent Pages deliver for demand measurement? +

Potent Pages designs and operates long-running crawling systems that convert public-web demand footprints into structured time-series datasets. Typical outputs include normalized tables, anomaly flags, and monitored recurring feeds.

Typical delivery: CSV, database tables, or API feeds with monitoring and change detection.
David Selden-Treiman, Director of Operations at Potent Pages.

David Selden-Treiman is Director of Operations and a project manager at Potent Pages. He specializes in custom web crawler development, website optimization, server management, web application development, and custom programming. Working at Potent Pages since 2012 and programming since 2003, David has extensive expertise solving problems using programming for dozens of clients. He also has extensive experience managing and optimizing servers, managing dozens of servers for both Potent Pages and other clients.

Web Crawlers

Data Collection

There is a lot of data you can collect with a web crawler. Often, xpaths will be the easiest way to identify that info. However, you may also need to deal with AJAX-based data.

Development

Deciding whether to build in-house or finding a contractor will depend on your skillset and requirements. If you do decide to hire, there are a number of considerations you'll want to take into account.

It's important to understand the lifecycle of a web crawler development project whomever you decide to hire.

Web Crawler Industries

There are a lot of uses of web crawlers across industries to generate strategic advantages and alpha. Industries benefiting from web crawlers include:

Building Your Own

If you're looking to build your own web crawler, we have the best tutorials for your preferred programming language: Java, Node, PHP, and Python. We also track tutorials for Apache Nutch, Cheerio, and Scrapy.

Legality of Web Crawlers

Web crawlers are generally legal if used properly and respectfully.

Hedge Funds & Custom Data

Custom Data For Hedge Funds

Developing and testing hypotheses is essential for hedge funds. Custom data can be one of the best tools to do this.

There are many types of custom data for hedge funds, as well as many ways to get it.

Implementation

There are many different types of financial firms that can benefit from custom data. These include macro hedge funds, as well as hedge funds with long, short, or long-short equity portfolios.

Leading Indicators

Developing leading indicators is essential for predicting movements in the equities markets. Custom data is a great way to help do this.

GPT & Web Crawlers

GPTs like GPT4 are an excellent addition to web crawlers. GPT4 is more capable than GPT3.5, but not as cost effective especially in a large-scale web crawling context.

There are a number of ways to use GPT3.5 & GPT 4 in web crawlers, but the most common use for us is data analysis. GPTs can also help address some of the issues with large-scale web crawling.

Scroll To Top