GET
/
integrations
/
web_crawler
/
index
JavaScript
import Hyperspell from 'hyperspell';

const client = new Hyperspell({
  apiKey: 'My API Key',
});

const response = await client.integrations.webCrawler.index({ url: 'url' });

console.log(response.resource_id);
{
  "source": "collections",
  "resource_id": "<string>",
  "status": "pending"
}

Authorizations

Authorization
string
header
required

API Key or JWT User Token. If using an API Key, set the X-As-User header to act as a specific user. A JWT User Token is always scoped to a specific user.

Query Parameters

url
string
required

The base URL of the website to crawl

max_depth
integer
default:2

Maximum depth of links to follow during crawling

Required range: 0 <= x <= 5
limit
integer
default:20

Maximum number of pages to crawl in total

Required range: 1 <= x <= 100

Response

Successful Response

source
enum<string>
required
Available options:
collections,
vault,
web_crawler,
notion,
slack,
google_calendar,
reddit,
box,
google_drive,
airtable,
algolia,
amplitude,
asana,
ashby,
bamboohr,
basecamp,
bubbles,
calendly,
confluence,
clickup,
datadog,
deel,
discord,
dropbox,
exa,
facebook,
front,
github,
gitlab,
google_docs,
google_mail,
google_sheet,
hubspot,
jira,
linear,
microsoft_teams,
mixpanel,
monday,
outlook,
perplexity,
rippling,
salesforce,
segment,
todoist,
twitter,
zoom
resource_id
string
required
status
enum<string>
required
Available options:
pending,
processing,
completed,
failed