Draft
Conversation
fix bug whereby read_files on index not loaded raised an error
- add domain filter - add to_zarr method - fix bug whereby coordinates were wrong when data interpolated on sdl
renamed 'id' to 'ARCOID'
- shapes are saved internally to avoid computation on every call - shapes is made compatible with a MassFetcher output and the peculiar 'ARCOID' which is used as _dummy_argo_uid
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.This suggestion is invalid because no changes were made to the code.Suggestions cannot be applied while the pull request is closed.Suggestions cannot be applied while viewing a subset of changes.Only one suggestion per line can be applied in a batch.Add this suggestion to a batch that can be applied as a single commit.Applying suggestions on deleted lines is not supported.You must change the existing code in this line in order to create a valid suggestion.Outdated suggestions cannot be applied.This suggestion has been applied or marked resolved.Suggestions cannot be applied from pending reviews.Suggestions cannot be applied on multi-line comments.Suggestions cannot be applied while the pull request is queued to merge.Suggestion cannot be applied right now. Please check back later.
This PR implements a naive approach to fetch core Argo data massively, i.e. to fetch a very large selection of measurements, as fast as possible.
This class should work as long as data fit in memory, so check out your RAM levels.
For reasonable performances, this is intended to be used with a local copy of the GDAC.
This class use a distributed.client.Client to scale Argo data fetching from a large collection of floats.
This class return data in argopy "Standard" user-mode, i.e. with:
Note
One month of global Argo data in standard user-mod is about 45Mb on disk as a zarr archive, so there is a lot of room for improvement in having ARCO (Analysis-Ready and Cloud-Optimized) Argo data with such approach.
Installation to play with this new feature
You need to install argopy from this branch.
This can be done like this:
Examples
Fetch and load the global Argo dataset for a given month
Fetch and save the global Argo dataset for a given month
Work with interpolated data
Load time series of global Argo dataset as points
Load time series of global Argo dataset as interpolated profiles