Cloudflare Developers

CD

Cloudflare Developers

Welcome to the official Cloudflare Developers server. Here you can ask for help and stay updated with the latest news

Join

hi, I'm reading gzip file produced by

hi, I'm reading gzip file produced by pipeline saved into R2 using boto3. But using s3.get_object on that object cause FlexibleChecksumError

Hello everyone,

Hello everyone, I am trying to follow along this tutorial: https://developers.cloudflare.com/pipelines/tutorials/query-data-with-motherduck/. I am stuck at the pipeline creation stage. When I run npx wrangler pipelines create clickstream-pipeline --r2-bucket clickstream-data --compression none --batch-max-seconds 5, I get the following error: ``` ⛅️ wrangler 4.27.0 (update available 4.28.0)...

We are using pipelines in two different

We are using pipelines in two different customer cloudflare accounts. Both pipelines are standard default deployments with a worker sending json data to the pipeline. In both cases pipelines ends up leaving aborted multipart data files in r2 which leads to permanent data loss. Anyone have any ideas?

Hi @kagitac I am trying to POC a switch

Hi @kagitac I am trying to POC a switch from kinesis firehose to CF pipelines, initially using the HTTP endpoint but also eventually moving from lambda as the entrypoint to a worker.... So far, I'm seeing pretty sluggish perf on ingestion, ranging from 400 to 1200ms, both when using the http api & a worker binding. With the worker binding im also seeing some Internal Operation Failed (which i've added retry for, but that makes the process even slower) Is this expected perf or could I be doing something funky? ...

Can we assign a custom URL to the

Can we assign a custom URL to the pipeline endpoint?

Hi Cole, yes sure:

Hi Cole, yes sure: bb5b7d2a19c4455faa46a808919b50aa thanks for the help!...

hi @Matt Silverlock, re: dynamic

hi @Matt Silverlock, re: dynamic partitioning – the documentation states "By default, Pipelines partition data by event date and time. This will be customizable in the future.". can you share any rough timelines for when you expect to ship this? pretty critical for multi tenant workloads, so looking to assess if i can give Pipelines a shot or go down another route 🙂 thx!...

Linking Cloudflare R2 as a source - Docs...

My goal is to store events for something like PostHog to injest via R2. Seems like csv and Parquet files format? https://posthog.com/docs/cdp/sources/r2 Would be great to collab on this!...

Whats the difference between pipelines

Whats the difference between pipelines and logpush exactly? Assuming a custom api vs the generic one controlled by logpush?

Is there any plans to go beyond 100MB

Is there any plans to go beyond 100MB for input files? Right now for some of my applications, I have to upload large files to AWS S3 first to calculate the MD5 and length, and then stream it to R2 after.