How It Works
- Load - Opens each URL in a real browser
- Execute - Runs your browser actions (clicks, scrolls, etc.)
- Solve - Automatically handles CAPTCHAs
- Process - Runs AI extraction (if prompt provided)
Browser Actions
Interact with pages before scraping:
{
"urls": [{
"url": "https://example.com/products",
"actions": [
{"type": "click", "selector": "#accept-cookies"},
{"type": "wait", "value": 1000},
{"type": "click", "selector": ".load-more-btn"},
{"type": "scroll", "value": "50%"}
]
}],
"prompt": "List all product names and prices",
"output": "json"
}
Available Actions
| Action | Description | Example |
|---|
click | Click an element | {"type": "click", "selector": "#btn"} |
type | Type into input | {"type": "type", "selector": "#search", "value": "query"} |
wait | Wait (ms) | {"type": "wait", "value": 2000} |
scroll | Scroll page | {"type": "scroll", "value": "50%"} |
Use browser DevTools (right-click → Inspect) to find CSS selectors.
Authentication (Optional)
Scrape protected pages by providing session cookies from your logged-in browser. We support two formats:
The traditional name=value format, separated by semicolons:
{
"urls": [{"url": "https://example.com/company/stripe"}],
"prompt": "Extract company details",
"output": "json",
"cookies": "authcookie=eyJ...; cf_clearance=2B08..."
}
Copy-paste directly from Chrome DevTools (Application → Cookies → Select All → Copy):
{
"urls": [{"url": "https://example.com/company"}],
"prompt": "Extract company details",
"output": "json",
"cookies": "authcookie\teyJhbGciOiJIUzUxMiJ9...\t.example.com\t/\t2026-06-30T14:29:30.522Z\t881\t✓\t✓\tLax\t\t\tMedium\ncf_clearance\tQnuFniylefl3k3FTfCbnp...\t.example.com\t/\t2027-01-01T14:29:32.591Z\t310\t✓\t✓\tNone\thttps://example.com\t\tMedium\n_ga\tGA1.1.1832229719.1766335524\t.example.com\t/\t2027-01-27T11:29:56.430Z\t30\t\t\t\t\t\tMedium"
}
The API auto-detects the format based on the presence of tabs and newlines. It extracts Name, Value, Domain, and Path from each row - extra columns (Size, HttpOnly, Secure, etc.) are ignored.
How to Get Cookies
- Log into the target website in your browser
- Open DevTools (F12) → Application → Cookies
- For Standard format: Copy individual cookie names and values, format as
name=value; name2=value2
- For Raw format: Select all rows (Ctrl/Cmd+A), copy (Ctrl/Cmd+C), and paste directly
Legal Responsibility: You are solely responsible for ensuring your authenticated scraping complies with applicable laws and the target website’s Terms of Service. Only scrape content you’re authorized to access. When uncertain, seek written permission from the platform. Misuse of session cookies may violate terms of service or applicable laws. Cookies are processed transiently and never stored by Spidra.
Array of URLs to scrape (1-3 URLs per request)
Required array length: 1 - 3 elements
Optional LLM prompt for extracting or transforming the scraped content
Output format for the extracted content
Available options:
json,
markdown
Enable stealth mode with proxy rotation to avoid detection
Available options:
queued
Unique job identifier for polling