crawlee-js
apify-platform
crawlee-python
💻hire-freelancers
🚀actor-promotion
💫feature-request
💻devs-and-apify
🗣general-chat
🎁giveaways
programming-memes
🌐apify-announcements
🕷crawlee-announcements
👥community
How to use Playwright's bypassCsp option?
Crawlee stops scanning for links with different anchors (#xyz) but the same base URL
How to access Actor input inside route handlers?
Connecting to a remote browser instance?
Crawlee Router as a folder with different files for each Handler
routes
in my project and then construct export const router
in index.ts
file by adding all my handlers from different files?
Like this
```typescript...Added "playwright-extra" with "stealthPlugin" and got error "Cannot read properties of undefined"
Module not found in NextJs projects
npm install crawlee playwright
but not puppeteer
Do i need puppeteer also to run playwright? i dont think so...
```...
Structure Crawlers to scrape multiple sites
site-1
and then site-2
initally before starting the crawler and then the crawler will dynamically add in the links as needed but this will mess up the logs since we are using a Queue and its FIFO, so first it'll crawl the first link, add the extracted links to the queue and then crawl the second link and its links to the queue and like this it'll keep switching contexts between the two sites which will make the logs a mess. Also routers, dont seem to have a url parameter, its just a category and then the request, so we will have to basically define handlers for each site in a single router right? which will just bloat up a single file.
Is there a better way I can structure this? usecase is to setup crawlers for 10+ sites and crawl them sequentially or in parallel but having sane logging for them....INSTAGRAM FOLLOWERS SCRAPER
Blocking requests after click
preNavigationHooks
to block images, which works for the initial page load but does not block images loaded after a click on the page (i.e. XHR requests). How can these be blocked?
```typescript
preNavigationHooks: [
async ({ page, blockRequests }) => {...relationship between memoryMbytes and availableMemoryRatio?
Help me!
How can I configure PuppeteerCrawler to not save request information to disk?
Hi, I'm new to Crawlee & Appify having discovered it yesterday
Can i get 403 status
not skipping over urls for unfound elements
Crawlee scrapper invoking the same handler multiple times
@crawlee/browser-pool useFingerprint + constant browser size
Scrape private website?