Ecosyste.ms: Awesome
An open API service indexing awesome lists of open source software.
https://github.com/hydroframe/hf_hydrodata
https://github.com/hydroframe/hf_hydrodata
Last synced: about 1 month ago
JSON representation
- Host: GitHub
- URL: https://github.com/hydroframe/hf_hydrodata
- Owner: hydroframe
- License: other
- Created: 2023-09-12T21:05:39.000Z (over 1 year ago)
- Default Branch: main
- Last Pushed: 2024-11-26T22:34:42.000Z (about 1 month ago)
- Last Synced: 2024-11-27T11:45:11.813Z (about 1 month ago)
- Language: Python
- Size: 4.58 MB
- Stars: 6
- Watchers: 6
- Forks: 1
- Open Issues: 1
-
Metadata Files:
- Readme: README.md
- License: LICENSE
Awesome Lists containing this project
- open-sustainable-technology - hf_hydrodata - This Python package is a product of the HydroFrame project and is designed to provide easy access to national hydrologic simulations generated using the National ParFlow model as well as a variety of other gridded model input datasets and point observations. (Hydrosphere / Ocean and Hydrology Data Access)
README
# hf_hydrodata
The ``hf_hydrodata`` Python package is a product of the [HydroFrame project](https://hydroframe.org) and is designed to provide easy access to national hydrologic simulations generated using the National ParFlow model ([ParFlow-CONUS1](https://hydroframe.org/parflow-conus1) and [ParFlow-CONUS2](https://hydroframe.org/parflow-conus2)) as well as a variety of other gridded model
input datasets and point observations. Some of the datasets provided here are direct observations
(e.g. USGS streamflow observations) while other are model outputs (e.g. ParFlow-CONUS2) or data products
(e.g. remote sensing products).[![DOI](https://joss.theoj.org/papers/10.21105/joss.06623/status.svg)](https://doi.org/10.21105/joss.06623)
## Installation
The best way to install `hf_hydrodata` is using pip. This installs our
latest stable release with fully-supported features:pip install hf_hydrodata
Users must create a HydroFrame API account and register their PIN before using the `hf_hydrodata` package. Please see [Creating a HydroFrame API Account](https://hf-hydrodata.readthedocs.io/en/latest/getting_started.html#creating-a-hydroframe-api-account) for detailed instructions.
## Documentation
You can view the full package documentation on [Read the Docs](https://hf-hydrodata.readthedocs.io).
Please see our [Python API Reference](https://hf-hydrodata.readthedocs.io/en/latest/api_reference.html) for detail on each core
method.## Usage
You can use `hf_hydrodata` to get access to both gridded and point observation data from various
datasets.You can view the available datasets and variables from [the documentation](https://hf-hydrodata.readthedocs.io)
or you can get the list of dataset and variables from functions.import hf_hydrodata as hf
datasets = hf.get_datasets()
variables = hf.get_variables({"dataset": "NLDAS2", "grid": "conus1"})You can get gridded data using the `get_gridded_data()` function.
import hf_hydrodata as hf
options = {
"dataset": "NLDAS2", "variable": "precipitation", "period": "hourly",
"start_time": "2005-10-1", "end_time": "2005-10-2", "grid_bounds": [100, 100, 200, 200]
}
data = hf.get_gridded_data(options)``hf_hydrodata`` supports access to a collection of site-level data from a variety of sources using the `get_point_data()` function.
The below syntax will return daily USGS streamflow data from January 1, 2022 through January 5, 2022
for sites that are within the bounding box with latitude bounds of (45, 50) and longitude bounds
of (-75, -50).from hf_hydrodata import get_point_data, get_point_metadata
data_df = get_point_data(
dataset = "usgs_nwis",
variable = "streamflow",
temporal_resolution = "daily",
aggregation = "mean",
date_start = "2022-01-01",
date_end = "2022-01-05",
latitude_range = (45, 50),
longitude_range = (-75, -50)
)
data_df.head(5)# Get the metadata about the sites with returned data
metadata_df = get_point_metadata(
dataset = "usgs_nwis",
variable = "streamflow",
temporal_resolution = "daily",
aggregation = "mean",
date_start = "2022-01-01",
date_end = "2022-01-05",
latitude_range = (45, 50),
longitude_range = (-75, -50)
)
metadata_df.head(5)Please see the [How To](https://hf-hydrodata.readthedocs.io/en/latest/point_data/index.html#how-to) section of our documentation for in-depth examples using the point module functions. Additionally, our team has developed the [subsettools](https://hydroframesubsettools.readthedocs.io/en/latest/) Python package which uses `hf_hydrodata` to access data and subsequently run a [ParFlow](https://parflow.readthedocs.io/en/latest/) simulation. Please see the `subsettools` documentation for full walk-through examples of extracting data for a domain and subsequently running a ParFlow simulation.
## State of the Field
The `hf_hydrodata` package spans multiple agencies, and includes both site-level observations and national gridded datasets. This allows users to interact with data from many sources with a single API call. Existing packages such as the [`dataRetrieval`](https://cran.r-project.org/web/packages/dataRetrieval/vignettes/dataRetrieval.html) R package provide some similar capabilities allowing users to access a breadth of hydrologic site-level surface water and groundwater observations from the USGS. However, the `dataRetreival` package is limited to USGS sources and is designed for R users. Our package goes beyond this to provide access to data from multiple agencies (for example the SNOTEL and FluxNet observation networks). The `hf_hydrodata` package provides a common syntax for acquiring such observations so that the user need not spend valuable research time learning multiple syntaxes to get all data relevant for their watershed. Additionally, the `hf_hydrodata` package provides users access to a wide selection of gridded data products. Many of these data products are not publicly available by other means including inputs and outputs from the national ParFlow model and multiple gridded atmospheric forcing datasets.## Build Instructions
To build the component you must have a Python virtual environment containing
the required components. Install the required components with:pip install -r requirements.txt
Edit the Python components in `src/hf_hydrodata` and the unit tests in `tests/hf_hydrodata` and the data catalog model CSV files in `src/hf_hydrodata/model`.
Use Excel to edit the CSV files so that files are saved in standard CSV format.Generate the documentation with:
cd docs
make htmlThis will validate the model CSV files and
generate the read-the-docs html into the html folder.## Testing
Our tests are located within the `tests/hf_hydrodata` directory of this repository. The full test suite is run automatically via Jenkins with each new Pull Request and subsequent commits. Jenkins executes the tests using `pytest` from the root directory.## License
Copyright © 2024 The Trustees of Princeton University and The Arizona Board of Regents on behalf of The University of Arizona, College of Science Hydrology & Atmospheric Sciences. All rights reserved.`hf_hydrodata` was created by William M. Hasling, Laura Condon, Reed Maxwell, George Artavanis, Will Lytle, Amy M. Johnson, Amy C. Defnet. It is licensed under the terms of the MIT license. For details, see the [LICENSE](https://github.com/hydroframe/hf_hydrodata/blob/main/LICENSE) file.
## Data Use Policy
The software is licenced under MIT licence, but the data is controlled by a [Data Use Policy](https://hf-hydrodata.readthedocs.io/en/latest/data_policy.html).## Report an Issue
If you have a question about our code or find an issue, please create a GitHub Issue with enough information for us to reproduce what you are seeing.## Contribute
If you would like to contribute to `hf_hydrodata`, please open a GitHub Issue with a description of your plan to initiate a conversation with our development team. Then detailed implementation review will be done via a Pull Request.