I’m getting some pressure to move to Postgres so some people can use the admin dashboard tools like Looker. Any chance D1 works with any of those things?
Is this not the reason that D1 cant compete with databases like Turso? I mean you would have to go via workers and a DO, which means theres a certain minimal latency. Is my understanding incorrect?
Question about D1 bindings to Pages. I know that when you bind D1 to a worker it automatically restricts which region the worker spins up to be as close to D1 as possible. Does it have the same behavior when you bind a pages project? If so, is there any way to get around that?
in theory if they had the same number of PoPs as us, and we had to go through Workers first and they didn't, yeah you'd gain 1-10ms of latency overhead on us
If it is 1 -10 ms latency, thats wonderful. Based on the earlier conversations we have had regarding about DO locations (where do DO live links), DO's are available only in a few DC (For example SGA for an Indian audience). And the minimal latency for DO location from MUmbai would be 200 ms according to the web site. I am making some assumptions here a. That the D1 replicas when they come into pic after Q1 would still live near SGA like locations due to colo reqs 2. Not just initial connections to DO, but subsequent latency for all communications would be the same. In that case, how can the latency be lower (like 10 ms) without smart placement? I would say having these many pops actually is a menace for performance as compared to bettering it. Anothe rpoint to ponder and I am thinking out aloud -> about a use case where I depend on an R2/KV and a DO ( i have such a use case) . I am going every where to collect my objects and the initial request would take multiple latency hits, I cache them using Cache API after the initial, but again theres no guarantee my next customer would hit the same pop.
I am currently developing a microservice-like project that processes ~2 TB of data from an (S)FTP server once a day (is D1 expensive - for this scale of data?).
This data should then be processed and filtered afterwards and written to a database. I want to manage the database with an ORM as well as migrations.
Has anyone already had experience with ORMs and migrations in the context of D1 and perhaps already implemented a small microservice project (or Wrangler for that matter)?
Oh, there really is such a thing. I thought something like that would be implemented with some proprietary packages.
But the migrations are implemented in SQL here, right? When I was still working a lot with Java, there was Liquibase or something like that. There you could define such migrations database-agnostically.
I think that would be really cool. I personally like D1 and Cloudflare. However, I write my applications for customers who sometimes have strong compliance regulations, such as hosting everything themselves, or only certain cloud providers are prescribed.
High-performance (around the globe) would be my concern. That's why I was thinking about D1.
Of course there are other larger cloud providers that offer something like geo-replicated storage - but I'm not sure how well they perform.
In fact, all data within these 2 TB is intended for public use. A web app that I would have implemented would then retrieve and display this data, for example.
It's basically a huge product database. Images separately in the CDN.
i would still recommend a non sql database for this. its much more easily horizontally scalable than SQL. ofc planetscale has achieved horizontal scalability with a mysql databases using vitess, but i dont think its enough for a 2 TB data.
Personally, I am a big fan of relational databases. I always assumed that they could be optimised well with the appropriate data structure.
Do you think NoSQL would make more sense for my use case? A few years ago when I looked into it, NoSQL was usually still behind relational databases. But that might have changed in the meantime ...
There are big updates 1-2 times a day because the data changes regularly. However, this would not be performance-critical. Only the reads are performance-critical.
Basically, I aggregate product data from different providers. As the descriptions are often in different languages, they are run through the Deepl API again and are to be written to a database. And a few more preprocessing steps to check how reputable the provider is... And the product images are uploaded to a blob storage or CDN to avoid relying on the provider's image service
In the end, this data should be displayed on a website where you can view and compare the active products.
Basically, a user will not be able to retrieve large amounts of data in one go. However, I would like to keep every single data record with very good performance.
A search function is also planned, but not necessarily the most important thing. I have already tried it with Meilisearch, but I had problems searching for a single ID, which is sometimes useful for my use case.
Just imagine a huge "products" table or bucket. An end user either clicks on a link that contains an ID or a "slug" and only wants the one product to be displayed.
Hello everyone, I have a question, I am sharing my cloudflare service with other friends, as a super administrator it allows me to use cloudflare Images but they ask my friends to pay again, do I have to give them any permission apart from cloud images?