Skip to content

nasa/earthdata-hashdiff

Repository files navigation

earthdata-hashdiff

Available on pypi

This repository contains functionality to read Earth science data file formats and hash the contents of those files into a JSON object. This enables the easy storage of a smaller artefact for tasks such as regression tests, while omitting metadata and data attributes that may change between test executions (such as timestamps in history attributes).

Features

While the following section describes the overall features of earthdata-hashdiff, further detailed examples can be found in docs/Using_earthdata-hashdiff.ipynb.

Generating hashed files

JSON files that contain SHA 256 hash values for all variables and groups in a netCDF4 or HDF-5 file can be generated using either the create_h5_hash_file or create_nc4_hash_file.

from earthdata_hashdiff import create_nc4_hash_file


create_nc4_hash_file('path/to/netcdf/file.nc4', 'path/to/output/hash.json')

The functions to create the hash files have two additional optional arguments:

  • skipped_metadata_attributes - this is a set of strings. When specified, the hashing functionality will not include metadata attributes with that exact name in the calculation of the hash for all variables or groups.
  • xarray_kwargs - this dictionary allows users to specify keyword arguments to xarray when the input file is opened as a dictionary of group objects. The default value for this kwarg is to turn off all xarray decoding for CF Conventions, coordinates, times and time deltas.

Skipping metadata attributes

Some metadata attributes of netCDF4 or HDF-5 files may vary based on when those files are generated. earthdata-hashdiff already omits the history and history_json metadata attributes of all groups and variables when constructing a hash. It is possible to specify further attributes to be omitted from the hash generation:

create_nc4_hash_file(
    'path/to/netcdf/file.nc4',
    'path/to/output/hash.json',
    skipped_metadata_attributes={'attribute_name_one', 'attribute_name_two'},
)

In the example above, neither of the values for metadata attributes with names attribute_name_one or attribute_name_two will be included in the calculation of a hash value for any variable or group in the input file.

Hashing GeoTIFF files

A similar JSON file can be created for a GeoTIFF file:

from earthdata_hashdiff import create_geotiff_hash_file

create_geotiff_hash_file('path/to/geotiff/file.tif', 'path/to/output/hash.json')

This function has one additional optional argument:

  • skipped_metadata_tags - this is a set of strings. When specified, the hashing functionality will not include GeoTIFF metadata tags with that name.

Comparisons against reference files

When a JSON file exists with hashed values, it can be used for comparisons. The public API provides h5_matches_reference_hash_file and nc4_matches_reference_hash_file, although these both are aliases for the same underlying functionality using xarray:

from earthdata_hashdiff import nc4_matches_reference_hash_file


assert nc4_matches_reference_hash_file(
    'path/to/netcdf/file.nc4',
    'path/to/json/with/hashes.json',
)

The comparison functions have three optional arguments:

  • skipped_variables_or_groups - the input for this kwarg is a set of string. The strings are the full paths to variables and groups, which tell the function to not check if the generated hash for those variables and groups are identical to the values in the JSON reference hash file. Note, the comparison function will still check that the input file contains the named variables and/or groups, even though it doesn't check their hashed value.
  • skipped_metadata_attributes - this set of strings, when specified, omits matching metadata attributes from the calculation of all variables and groups. If metadata attributes were specified as skipped when generating the JSON file containing hashes, the same metadata attributes will need to be specified as skipped during comparison, to ensure the hashes match.
  • xarray_kwargs - this dictionary allows users to specify keyword arguments to xarray when the input file is opened as a dictionary of group objects. The default value for this kwarg is to turn off all xarray decoding for CF Conventions, coordinates, times and time deltas.

Omitting metadata attributes

If metadata attributes were omitted from hash calculations with create_nc4_hash_file or create_h5_hash_file, those same metadata attributes will need to be omitted from the comparison assertion.

assert nc4_matches_reference_hash_file(
    'path/to/netcdf/file.nc4',
    'path/to/json/with/hashes.json',
    skipped_metadata_attributes={'attribute_name_one', 'attribute_name_two'},
)

Comparisons with GeoTIFFs

The same operation can also be performed for a GeoTIFF file in comparison to an appropriate JSON reference file:

from earthdata_hashdiff import geotiff_matches_reference_hash_file

assert geotiff_matches_reference_hash_file(
    'path/to/geotiff/file.tif',
    'path/to/json/with/hash.json',
)

A single entry point for comparison

For convenience, you can use the matches_reference_hash_file for all of the file types previously discussed. Each call will accept the paths to the binary file and JSON hash file, along with appropriate optional kwargs relevant to the file type.

from earthdata_hashdiff import matches_reference_hash_file

assert matches_reference_hash_file(
    'path/to/netcdf/file.nc4',
    'path/to/json/with/hashes.json',
)

assert matches_reference_hash_file(
    'path/to/netcdf/file.nc4',
    'path/to/json/with/hashes.json',
    skipped_metadata_attributes={'attribute_name_one', 'attribute_name_two'},
)

assert geotiff_matches_reference_hash_file(
    'path/to/geotiff/file.tif',
    'path/to/json/with/hash.json',
)

assert geotiff_matches_reference_hash_file(
    'path/to/geotiff/file.tif',
    'path/to/json/with/hash.json',
    skipped_metadata_tags={'tag_name_one'},
)

Installing

Using pip

Install the latest version of the package from PyPI using pip:

$ pip install earthdata-hashdiff

Other methods:

For local development, it is possible to clone the repository and then install the version being developed in editable mode:

$ git clone https://github.com/nasa/earthdata-hashdiff
$ cd earthdata-hashdiff
$ pip install -e .

Developing

Development within this repository should occur on a feature branch. Pull Requests (PRs) are created with a target of the main branch before being reviewed and merged.

Releases are created when a feature branch is merged to main and that branch also contains an update to the earthdata_hashdiff.__about__.py file.

Development Setup:

Prerequisites:

  • Python 3.11+, ideally installed in a virtual environment, such as pyenv or conda.
  • A local copy of this repository.

As an example to set up a conda virtual environment:

conda create --name earthdata-hashdiff python=3.12 --channel conda-forge \
    --override-channels -y
conda activate earthdata-hashdiff

Install dependencies:

pip install -r requirements.txt -r dev-requirements.txt -r tests/test_requirements.txt

Running tests

earthdata-hashdiff uses pytest to execute tests. Once test requirements have been installed via pip, you can execute the tests:

pytest tests

The CI/CD workflows that execute the tests also make use of pytest plugins to additionally create code test coverage reports and JUnit XML output. These extra outputs can be produced with the following command:

pytest tests --junitxml=tests/reports/earthdata-hashdiff_junit.xml \
    --cov earthdata_hashdiff --cov-report html:tests/coverage --cov-report term

This will produce:

  • The test results (pass/fail) in the terminal.
  • A coverage report in the terminal running the tests. The coverage report will cover the contents within the earthdata_hashdiff directory.
  • An HTML format coverage report in the tests/coverage directory. The entry point for this output is tests/coverage/index.html.
  • JUnit style output in tests/reports/earthdata-hashdiff_junit.xml.

pre-commit hooks

This repository uses pre-commit to enable pre-commit checks that enforce coding standard best practices. These include:

  • Removing trailing whitespaces.
  • Removing blank lines at the end of a file.
  • Ensure JSON files have valid formats.
  • ruff Python linting checks.
  • black Python code formatting checks.
  • mypy Type hint checking and enforcement.

To enable these checks locally:

# Install pre-commit Python package:
pip install pre-commit

# Install the git hook scripts:
pre-commit install

Versioning

Releases for the earthdata-hashdiff adhere to semantic version numbers: major.minor.patch.

  • Major increments: These are non-backwards compatible API changes.
  • Minor increments: These are backwards compatible API changes.
  • Patch increments: These updates do not affect the API to the service.

Contibuting

Contributions are welcome! For more information see CONTRIBUTING.md.

About

No description, website, or topics provided.

Resources

License

Contributing

Stars

Watchers

Forks

Packages

No packages published

Contributors 3

  •  
  •  
  •