crawlee-js
apify-platform
crawlee-python
💻hire-freelancers
🚀actor-promotion
💫feature-request
💻devs-and-apify
🗣general-chat
🎁giveaways
programming-memes
🌐apify-announcements
🕷crawlee-announcements
👥community
Scrape data from TikTok for research
I want webhooks to send input url in the post request
Is there a way to initiate crawlee crawl + scraping jobs from a server?
application/octect stream in cheerio
ERROR CheerioCrawler: Request failed and reached maximum retries. Error: Resource http://127.0.0.1/website.web/part_to_scrape served Content-Type application/octet-stream, but only text/html, text/xml, application/xhtml+xml, application/xml, application/json are allowed. Skipping resource.
ERROR CheerioCrawler: Request failed and reached maximum retries. Error: Resource http://127.0.0.1/website.web/part_to_scrape served Content-Type application/octet-stream, but only text/html, text/xml, application/xhtml+xml, application/xml, application/json are allowed. Skipping resource.
Cheerio memory error
WARN CheerioCrawler:AutoscaledPool:Snapshotter: Memory is critically overloaded. Using 1174 MB of 750 MB (157%). Consider increasing available memory.
WARN CheerioCrawler:AutoscaledPool:Snapshotter: Memory is critically overloaded. Using 1174 MB of 750 MB (157%). Consider increasing available memory.
mixed headful and headless in a PlaywrightCrawler
I found that Apify took about 45$ from my credit for nothing
Facebook events by page
Firefox Error in PlaywrightCrawler
w9I8udSOta4b0kEw8
w9I8udSOta4b0kEw8
Is it possible to add request in middle of queue?
CheerioCrawler Timeout after 320 Seconds Error/Exception
2023-06-08T07:28:54.464Z ERROR CheerioCrawler: An exception occurred during handling of failed request. This places the crawler and its underlying storages into an unknown state and crawling will be terminated. This may have happened due to an internal error of Apify's API or due to a misconfigured crawler.
2023-06-08T07:28:54.464Z ERROR CheerioCrawler: An exception occurred during handling of failed request. This places the crawler and its underlying storages into an unknown state and crawling will be terminated. This may have happened due to an internal error of Apify's API or due to a misconfigured crawler.
Can't mark request as failed
request.pushErrorMessage()
method. The session.retire()
is working for marking the session as bad, but not for mark the request as failed:
...Keyvaluestore file extensions
.mhtml
file extension? Using the code below seems to always set it to .bin
extension
```...Pause concurrent requests ?
web scrapper create 1 file instead of multiple output for pagination
apify run
I get 20 different json files. How can I combine all the data in 1 json file for all the pages?Error when running in Docker Container
node src/main.js...
conducting faster scrapes with pagination and individual product scraping
Crawlee + Proxy = Blocked, My laptop + Proxy = unblocked
double import problem