Firecrawl

F

Firecrawl

Join the community to ask questions about Firecrawl and get answers from other members.

Join

❓┃community-help

💬┃general

🛠️┃self-hosting

ERROR: 408 on Scrape

Request failed with status code 408 at FirecrawlApp.scrapeUrl (webpack-internal:///(rsc)/./node_modules/@mendable/firecrawl-js/dist/index.js:78:13) Any ideas on what this error code means?...

Bulk Crawling

Hello, In my app, users enter a URL and select additional links from a list returned by the /map endpoint that they would like to crawl alongside the main URL they provided. Currently, I'm sending a request to the /crawl endpoint with the main URL being the one the user provided, and the includePaths:[] containing the paths the user selected from the /map endpoint. However, it seems that all the links from the main URL are being crawled, in addition to the selected paths. How can I modify my request so that only the main URL and its selected includePaths are crawled, excluding any other links?...

Reduced /map endpoint results

Till yesterday, /map endpoint was returning 5000 links but now it’s only 99 links. Did it get reduced or a bug ?...

base64 images

When crawling some pages, it picks up on base64 images, making it not fun to put that into vector stores 😄 any suggestions?

Does map_url support external links?

I tried using allowExternalLinks in map_url, but none of the external links appear. Am i missing something? Or is this not supported yet?

http 👉 https

Let's say I'm sending a url starting with http://, the url redirects to https://. Will sublings starting with https:// be crawled?

408 errors on both API and Dashboard

Hey guys, are y'all experiencing issues? I cannot scrape anything pages currently. Thank you!

I want the main content and all the css + js links in the html. How do I do it?

I attempt to set: 'formats': ['html', 'rawHtml'], 'onlyMainContent': True and hope that the rawHtml will be the html + (css + js) files. But it is not true. It is the original html without noise content filtering. ...

`next` pagination using js-sdk

Using version 1.2.2, does the js-sdk have a built-in method to follow the next url? I call asyncCrawlUrl() and then checkCrawlStatus() every 5 seconds until the job status is complete. However, I'm not sure how to get the next page of results from the next property. More detailed examples of how to use asyncCrawlUrl, checkCrawlStatus, and the next url would be appreciated!...

Limit speed each request

Hi, I dont know how the crawl works but it's possible to slow the crawl and dont get detected ? I check the doc : limit is the total page crawl poll_interval is the status of the crawl ...

Cannout get around of cloudfare with the nodejs API call

This is my code const params = { pageOptions: { includeHtml: true,...
No description

Unexpected error happens with the scrape of https://nivod.cc/ using curl

this is the response {"success":false,"error":"An unexpected error occurred. Please contact hello@firecrawl.com for help. Your exception ID is 9b0fd50d-7f45-4dbb-9af5-5d81b39ca5f8"}y...

Scraping websites built with GoDaddy

Hi, I am using Firecrawl and I have some issues with scraping websites built with GoDaddy, can some team member contact me so I can send more examples of failed attempts? https://leadtogetherllc.com/ Using v0, v1 and your Playground (v1), both /scrape and /crawl endpoints don't work....
No description

stuck with python sdk, works on and off with curl

I just selfhosted a firecrawl instance yesterday, it works immediately with a curl request to /v1/crawl, which is a surprise! and then.... 1. with curl i built playwright-ts and it doesn't work at first because i didn't really think when I quoted PLAYWRIGHT_MICROSERVICE_URL=http://localhost:3000/scrape from the SELF_HOST.md into my envfile, which is obviously incorrect. after I fixed it i works with example request ```...

'limit' not respected?

Hi there - I'm having trouble getting the limit option respected; this was true in v0 and now in v1. I have just built the docker image from the main repo and running local. Different permutations of limit and maxDepth return different results, but not in a way I understand. ...