Zarr dataset on requester pays bucket only opening if I list the files first!?
See original GitHub issue@martindurant , this is pretty bizarre: I have a notebook that used to open a zarr dataset (requester pays bucket), but after upgrading my environment, opening the dataset no longer works unless I list the files first!?
So this doesn’t work:
xr.open_zarr(fsspec.get_mapper('s3://pangeo-data-uswest2/esip/adcirc/adcirc_01d',
anon=False, requester_pays=True))
but this does:
fs = fsspec.filesystem('s3',anon=False, requester_pays=True)
fs.ls('s3://pangeo-data-uswest2/esip/adcirc/adcirc_01d/')
xr.open_zarr(fsspec.get_mapper('s3://pangeo-data-uswest2/esip/adcirc/adcirc_01d',
anon=False, requester_pays=True)
Here’s the notebook: https://nbviewer.jupyter.org/github/rsignell-usgs/hurricane-ike-water-levels/blob/master/fsspec_strangeness.ipynb
I don’t have a minimal test, but it should be reproducible on binder using this link: and running the “fsspec_strangeness” notebook.
This is using fsspec=0.8.0
and xarray from master
. I guess this could also be an xarray issue?
I have no idea. 😕
But I bet you do. 😄
Issue Analytics
- State:
- Created 3 years ago
- Comments:9 (8 by maintainers)
Top Results From Across the Web
Zarr dataset on requester pays bucket only opening if I list the ...
I suspect that the bucket has list_object permissions, but not get_object permissions. If the files have not already been listed (so their details...
Read more >Convenience functions (zarr.convenience) — zarr 2.13.3 ...
Convenience function to open a group or array using file-mode-like semantics. Parameters ... Store or path to directory in file system or name...
Read more >zarr-developers/community - Gitter
Hi I have a question about the zarr files served via http. In the zarr files, the chunks are stored in separate files...
Read more >Cleaning out the pangeo-data google cloud storage bucket
Sorry @rabernat, there seems to be an issue with the pangeo-meom bucket, I get the error “BadRequestException: 400 Bucket is requester pays ......
Read more >Pangeo Forge: Crowdsourcing Analysis-Ready, Cloud ...
Just like ARCO data production, installing open source software ... one or many NetCDF files into a single consolidated Zarr dataset.
Read more >
Top Related Medium Post
No results found
Top Related StackOverflow Question
No results found
Troubleshoot Live Code
Lightrun enables developers to add logs, metrics and snapshots to live code - no restarts or redeploys required.
Start Free
Top Related Reddit Thread
No results found
Top Related Hackernoon Post
No results found
Top Related Tweet
No results found
Top Related Dev.to Post
No results found
Top Related Hashnode Post
No results found
See https://github.com/aio-libs/aiobotocore/issues/825
You can try setting the environment variable AWS_DEFAULT_REGION as a workaround, or adding
client_kwargs={"default_region": ...}
intoget_mapper
Yep - this is something that aiobotocore needs to wort out for us - it would be good to chime in on that issue to give your specifics, to see if it can help them correct the problem. https://github.com/aio-libs/aiobotocore/issues/825