Replies: 2 comments 4 replies
-
|
hi @timonmerk, there are a lot of ways to copy zarr data; where are you transferring the data from, and where are you transferring it to? Are you sure the slowness has anything to do with zarr itself, and not your network connection? |
Beta Was this translation helpful? Give feedback.
3 replies
-
|
does the consumer need the entire thing, or just part of it? hosting the dataset over http from HPC 1, and having the consumer on HPC 2 access (and potentially cache) just the parts they need could be an alternative to globus. |
Beta Was this translation helpful? Give feedback.
1 reply
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Uh oh!
There was an error while loading. Please reload this page.
-
I need to transfer a zarr dataset, and the transfer throughput is unfortunately prohibitively slow. This is likely due to the many files resulting from smaller chunk size that is optimal in our downstream application.
Currently I am using tar to compress individual blocks that makes the transfer faster. I am not sure though if this is the optimal solution, since it requires additional manual steps.
Unfortunately rechunking might break a lot of our chunk-associated metadata.
Beta Was this translation helpful? Give feedback.
All reactions