Best Firecrawl methods for large recurring tasks

I have another question if you @Gaurav Chadha don't mind.
Without addressing the appropriateness of such tasks, I'd like to do the following using Firecrawl -- 1. do a search for all bid publishing sites for one service 2. do a scrape of each site, extracting bids with a "Current" or "Open" criteria (dealing with pagination if present) 3. do a scrape of each bid, extracting a known schema for each bid What is the most cost effective way to do this? Doing step 1 every week, and steps 2 and 3 every day. It seams the extract method is more expensive, but is it cheaper for large jobs. Steps 2 and 3 are pretty "deterministic" without using an AI model, but when is using AI worth the extra cost?
1 Reply
Gaurav Chadha
Gaurav Chadha2w ago
Hi @Robin Mattern you can use cron job instead of doing these tasks each week to be more efficient and avoid any manual work. refer to the #3 - https://www.firecrawl.dev/blog/n8n-web-scraping-workflow-templates example for step1 - use /search query with category for step 2 and 3 /scrape as discussed here - https://discord.com/channels/1226707384710332458/1418585477342167072 yeah, extract is much better for larger jobs and as it comes with AI LLM it's much more efficient, and it comes with 500k token initially https://docs.firecrawl.dev/features/extract
Firecrawl
Web Scraping with n8n: 8 Powerful Workflow Templates
Learn how to build robust web scraping automations with n8n workflow templates that extract data from any website without coding using Firecrawl's AI-powered engine.
Firecrawl Docs
Extract | Firecrawl
Extract structured data from pages using LLMs

Did you find this page helpful?