How to use the fastCRW crawl flow for asynchronous multi-page extraction and job polling.
Use crawl when you need multiple pages instead of a single response payload. It is the right tool for:
curl -X POST https://fastcrw.com/api/v1/crawl \
-H "Content-Type: application/json" \
-H "Authorization: Bearer YOUR_API_KEY" \
-d '{"url":"https://example.com","limit":50}'
The initial call returns a job identifier. Poll the job endpoint until the crawl is complete.
The crawl API is asynchronous by design.
POST /crawl starts the job.GET /crawl/:id returns progress and newly available results.completed or a terminal error is returned.curl https://fastcrw.com/api/v1/crawl/CRAWL_ID \
-H "Authorization: Bearer YOUR_API_KEY"
That flow is easy to drive from shell scripts, job runners, background workers, and dashboards.
The exact crawl body can vary by workflow, but the most common fields are:
| Field | Description |
|---|---|
url | Required starting URL |
limit | Maximum number of pages to collect |
maxPages / max_pages | Supported aliases for crawl caps during compatibility-oriented migrations |
Start small. A crawl with limit: 5 is much easier to inspect than a crawl with limit: 500.
The safest way to evaluate a new site is:
map first to understand the reachable section,That sequence saves credits and helps you catch bad starting URLs early.
crawl billing is different from scrape because results materialize over time.
If the API returns 429, respect Retry-After. If the target site itself is slow or hostile, reducing crawl size usually gives you a clearer signal than hammering the same job harder.
The polling model is explicit today. That keeps the API easy to understand from scripts, pipelines, and dashboard tooling. It also makes billing and progress reporting easier to reason about than hidden background behavior.