-
Notifications
You must be signed in to change notification settings - Fork 27
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Setting rechunk_max_mem to worker memory doesn't work #54
Comments
That's a strange one. It's likely an issue in distributed rather than rechunker. What version of dask are you using? I'm not familiar with how SLURMCluster handles environments, but is the version "on the cluster" the same as your client ( |
|
@TomAugspurger , well, you were right, there is something funky with my use of SlurmCluster. When I tried a LocalCluster it worked fine. |
@TomAugspurger , and now I've discovered the real problem. It wasn't a problem with So specifying |
I think that For this workload (where the memory usage can be very tightly controlled), the best option might be setting You'll want at least some buffer for other stuff in the Python process (like module objects). (If that's correct, then we should document it 😄) |
For a worker memory of 6GB, I tried systematically reducing the |
My first time using rechunker!
I'm running on HPC with a SlurmCluster, reading and writing zarr to fast local disk.
This runs for about a minute, then barfs with
keyerror: startstops
:https://nbviewer.jupyter.org/gist/rsignell-usgs/fe0a1ec0cec562b14ddcc9a222ddc34f
The text was updated successfully, but these errors were encountered: