Crawl status tool is returning 500, need help to verify whats wrong with the request.
I am calling firecrawl_check_crawl_status tool in remote mcp with the following argument. {"id": job_id}. Acticity logs shows just one page being crawled although I set depth 2, limit 3. Posted a screenshot of activity logs (showing a different jobid). I have set a depth and limit but the log says one one page is crawl and 5 credit is used.
Response:
{'result': {'content': [{'type': 'text', 'text': 'Error: Failed to check crawl status. Status code: 500. Error: An unexpected error occurred. Please contact help@firecrawl.com for help. Your exception ID is 59b6e23c4a03481fa61b958c7b08c15c'}], 'isError': True}, 'jsonrpc': '2.0', 'id': 1754498520281}

11 Replies
I also see both scrape and crawl job being called.

Hey there! So the duplicate crawl and scrapes in your activity log is a known bug (introduced about 2 weeks ago when we were making performance improvements to the dashboard). We will try to get that fixed soon.
The reason that this request cost 5 credits per page instead of 1 credit per page is because you passed the "json" format into the change tracking feature. I'd recommend not passing "json" there since you aren't passing in the jsonOptions scraping parameter anyway.
And for that exception ID, it looks like a badly formatted string was passed into the request. Here's the error I see on my side:
invalid input syntax for type uuid: "3ccbe36a-09ca-45e1-b6ca-29aa757fe88c. Use firecrawl_check_crawl_status to check progress."
thank you. I added the fixes you suggested. I am struck at something else now. I can see the crawl status complete. But the result is truncated. How to get the full result given a job id?
We return 10MB of data at a time. To get the next page, you can check the
next
parameter: https://docs.firecrawl.dev/api-reference/endpoint/crawl-get#response-next.
There's a chance that n8n might be truncating the results on that side though? Not sure how much context it lets you work with.I tried the endpoints and I still dont see the data. {
"success": true,
"status": "completed",
"completed": 0,
"total": 0,
"creditsUsed": 0,
"expiresAt": "2025-08-07T01:36:38.000Z",
"data": []
}
So it looks like the crawl parameters you are passing are preventing any pages from being crawled. For example, you'll want to set maxDepth to something higher than 2. The default of 10 would be better here. And I'd remove the includePaths parameter, since what you passed for that regular expression doesn't match any of the URLs on the site.
it seems to work fine in playground, thats what is confusing me. \
Can you share the working playground link with me?
It will be greatly helpful if you you explain how include paths is supposed to work. Lets say I have this link: https://www.weforum.org/stories/economic-growth/ and I want to crawl to each of the articles in that link. This is one such article. https://www.weforum.org/stories/2025/07/globalization-multinational-corporations/. so in this case if I do includepath "/2025/*" is it supposed to grab all articles under it? I am trying to prevent the crawler to going off track.
Thanks for sharing that example! For that you'd want to set includePaths to
stories/2025/.*
and also set crawlEntireDomain=true. Here's a working playground link.
Hey there,
It's been a while since we've last heard from you, so we are going to go ahead and close out this ticket.
Feel free to open a new one, as needed!thank you