It seems my bucket has ground to a near

It seems my bucket has ground to a near halt. I was previously uploading a 130MiB/s+ and now I'm seeing Completed 312.0 MiB/4.7 GiB (4.1 MiB/s) with 1 file(s) remaining from the aws cli. df681cd207a9e2afa394d260c486fd1e bucket tts-data
D
DanTheGoodmanā€¢43d ago
I have a handfull of multipart uploads I abanadoned, not sure if that matters but with 7.8T to upload, this is a major problem lol it seems to be going fast again after I deleted the few files that were in there interesting, will usage that in the future. Should I use sync command or something else? pinging reply šŸ™‚ gotcha, thanks will try that! do you know if it can resume partial uploads @Space - Ping in replies ? šŸ˜¦ well s3 has resumable multipart uploads no? Because copy skips existing files, if I cancel the current aws s3 cp command it will resume where it left off, minus the abandoned upload? Copy files from source to dest, skipping identical files. sure, I mean like if I have file abc already uploaded to dest and it's on disk, when I run rclone copy it will skip copying that bc it sees it is already in dest right? ok cool that's great then I can just kill my aws scripts haha that'd be great! im moving from local to r2 actually haha oh duh you said that will this make a noticeable improvement over default settings @Space - Ping in replies ? im seeing just a ton of 0%'s
* da_2019_2.tar: 0% /17.562Gi, 0/s, -
* da_2020_2.tar: 0% /3.340Gi, 0/s, -
* de_2009.tar: 0% /6.347Gi, 0/s, -
* de_2009_2.tar: 0% /13.354Gi, 0/s, -
* de_2010.tar: 0% /8.340Gi, 0/s, -
* de_2011.tar: 0% /6.984Gi, 0/s, -
* de_2011_2.tar: 0% /31.641Gi, 0/s, -
* de_2012_2.tar: 0% /30.311Gi, 0/s, -
* de_2013.tar: 0% /10.115Gi, 0/s, -
* de_2014.tar: 0% /4.800Gi, 0/s, -
* de_2014_2.tar: 0% /19.833Gi, 0/s, -
* de_2015_2.tar: 0% /32.594Gi, 0/s, -
* da_2019_2.tar: 0% /17.562Gi, 0/s, -
* da_2020_2.tar: 0% /3.340Gi, 0/s, -
* de_2009.tar: 0% /6.347Gi, 0/s, -
* de_2009_2.tar: 0% /13.354Gi, 0/s, -
* de_2010.tar: 0% /8.340Gi, 0/s, -
* de_2011.tar: 0% /6.984Gi, 0/s, -
* de_2011_2.tar: 0% /31.641Gi, 0/s, -
* de_2012_2.tar: 0% /30.311Gi, 0/s, -
* de_2013.tar: 0% /10.115Gi, 0/s, -
* de_2014.tar: 0% /4.800Gi, 0/s, -
* de_2014_2.tar: 0% /19.833Gi, 0/s, -
* de_2015_2.tar: 0% /32.594Gi, 0/s, -
like 500 of these lol preallocating where? can I do that without writing the extra files? I don't want them there if they don't have to be
on r2 and making sure to write extra metadata hashes etc
what does this mean then? does this process take a while? oh wow ok well that might not end up being faster lol
Transferred: 0 B / 7.771 TiB, 0%, 0 B/s, ETA -
Checks: 159 / 159, 100%
Transferred: 0 / 550, 0%
Elapsed time: 4m44.5s
Transferring:
* bg_2009_2.tar: 0% /13.328Gi, 0/s, -
* bg_2009.tar: 0% /5.104Gi, 0/s, -
* bg_2010.tar: 0% /8.276Gi, 0/s, -
* bg_2010_2.tar: 0% /33.142Gi, 0/s, -
Transferred: 0 B / 7.771 TiB, 0%, 0 B/s, ETA -
Checks: 159 / 159, 100%
Transferred: 0 / 550, 0%
Elapsed time: 4m44.5s
Transferring:
* bg_2009_2.tar: 0% /13.328Gi, 0/s, -
* bg_2009.tar: 0% /5.104Gi, 0/s, -
* bg_2010.tar: 0% /8.276Gi, 0/s, -
* bg_2010_2.tar: 0% /33.142Gi, 0/s, -
says it's checked 159 though? I don't have too much disk left though lol Is there progress on that reading I can show? This isn't a very fast disk (network volume) I also added the no_check_bucket = true option to the config because fo the keys no it's a network volume on DO sorry, like a block storage not locally attached storage, but still a "disk" digitalocean magic no it shows like a local disk it's just not as fast as a locally attached disk I just wish I could see the progress of the preallocation, it's been going for 9 min and no % changes lol will the preallocation need to restart though? I guess so, since it shows 4 now lol. I guess it preallocates 4 at a time then? oh yeah there goes the first one @Space - Ping in replies it only looks like it's doing 1 file at a time though
Transferred: 7.907 GiB / 7.771 TiB, 0%, 162.067 MiB/s, ETA 13h57m6s
Checks: 159 / 159, 100%
Transferred: 1 / 550, 0%
Elapsed time: 1m35.0s
Transferring:
* bg_2009_2.tar: 0% /13.328Gi, 0/s, -
* bg_2010.tar: 33% /8.276Gi, 113.020Mi/s, 49s
* bg_2010_2.tar: 0% /33.142Gi, 0/s, -
* bg_2011.tar: 0% /6.799Gi, 0/s, -
Transferred: 7.907 GiB / 7.771 TiB, 0%, 162.067 MiB/s, ETA 13h57m6s
Checks: 159 / 159, 100%
Transferred: 1 / 550, 0%
Elapsed time: 1m35.0s
Transferring:
* bg_2009_2.tar: 0% /13.328Gi, 0/s, -
* bg_2010.tar: 33% /8.276Gi, 113.020Mi/s, 49s
* bg_2010_2.tar: 0% /33.142Gi, 0/s, -
* bg_2011.tar: 0% /6.799Gi, 0/s, -
and this is at 4-10gbit machine oh ok now some are starting to run concurrently, maybe this disk is just a lot slower than I thought lol the single files don't seem to be uploading faster than just dumb concurrent uploads from aws s3 cp though, those ran ~110MiB/s each yeah I think this is slower because of all the file reading it has to do first hahaha damn well not if it's limited by disk speed instead of network right? but it has to read every file twice effectively with rclone, is there a way to do a list of web files directly to S3? R2 using DO just as a download/upload machine since bandwidth is cheaper and I can get "10gbit" machine which really seems to be 4 lol I'm trying to download a 7.7TB data set and upload into my own R2 some web server with a list of direct tar files like I have the list of links rclone copyurl? I don't think the server provides file listings though, but I have them in a txt file sec https://dl.fbaipublicfiles.com/voxpopuli/audios/ they have code to generate the list of links but I don't think I can do them concurrently with a single rclone command, I'd have to run a bunch in parallel issue is if that breaks, it won't know to skip already uploaded ones right? I don't see on the docs for that page that it would ah see that now ill test somethin @Space - Ping in replies I was getting 400MiB/s earlier, but sometimes get like 50 You think I could launch 552 of these concurrently? Or would that be dumb? I don't think there's an easy way to do like 10 concurrently without writing code to manage it let me try that honestly that's fine, I've already been messing up for 2 days XD that way I can save $1/hr on not having a 10TB disk XD wdym? LMAO christ how are you able to saturate even close to that from the source with concurrent downloads? I don't want to bum your resources, I'm making a new node now to try running this let me see hwo fast I can get this one node to trasnfer real quick and if it's slow I'll send you the details, just made a max size hetzner VM lol
Want results from more Discord servers?
Add your server
More Posts
Number of whole builds?Where can I see the number of whole builds? I want to know whether the limit has been reached šŸ˜†Load Balancer - Mystery 404sHi all, I have created a CloudFlare load balancer: https://ml-inference-lb.photoninsights.com/ It poWARP on macOS - memory leak?I noticed yesterday that the `CloudflareWARP` process had grown to more than 4GB on my laptop. Afteany chance we'll see something like thisany chance we'll see something like this for the models cloudflare uses?Getting Internal Server Error from Workers AIThis issue happens when a Directory or File Path is included in the prompt for `@hf/thebloke/deepseecloudfare never working on my oculus quest twotryna sign into nexus mods on my q2 but it always failcan i host .net framework on cloudflare pages ?can i host .net framework on cloudflare pages ?can i host .net framework on cloudflare pages ?can i host .net framework on cloudflare pages ?Created Default DNS Location by Accident on Zero TrustI was exploring the zero trust options in the cloudflare dashboard and accidentally created a DNS loHow to add "Content-Encoding: br" to Cloudflare R2 StorageI'm trying to serve a Unity WebGL Build compressed to Brotli from R2 Storage but I have no idea how Cloudflare's secrets functionality & Mailjet?Trying to authenticate with Mailjet and finding send mail errors. API key and secret are correct andWhere to find API endpoints to create internal/external hostnames?I have a Cloudflare tunnel setup to my home network for some personal self-hosted services. While lethird party github option for cloudflare pagesIs there a way I could have cf pages pull from forgejo?After clearing all caches, the svg images of 3 URLs cannot be opened.After clicking to clear all caches on cloudflare, the svg images of 3 URLs cannot be opened. Does anHelp setting up worker for discord botHello, I wanted to host a discord bot that embeds tiktok links in my friends server. I found this reRedirect not working when transforming via Workers onError eventHi folks, I have a worker that resizes images; everything works well except when the animated GIF ican i run --force here ?can i run --force here ?Workers KV get throws TypeError: Cannot read properties of undefined (reading 'get')My wrangler.toml file looks like this: ``` kv_namespaces = [ { binding = "BH_API_KEYS", id = "xxxPaid for the Services.Paid for support, my site is under DDOs attacks I would like a dev to help me resolve this issue. passh over cloudlare tunnel with iPhoneHi, I successfully got an ssh connection through a cloudflare tunnel on windows but it worked becaus