-
Couldn't load subscription status.
- Fork 14
Specify files instead of globbing, enable logging, add testing #54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Specify files instead of globbing, enable logging, add testing #54
Conversation
|
@aerorahul Creating a |
|
Installing grib2io on ursa needs g2c. Discussed with @LinlinCui-NOAA and got that resolved. Since this is using |
This PR includes the last changes that SPA team made to - `oper/gen_aigfs_ics.py` - `oper/gen_aigefs_ics.py` to operationalize the scripts. These changes are a band-aid fix that will need a coordinated fix in the next release. A PR such as #54 would likely resolve these issues. Some key areas we will need to improve on include: - `oper/gen_aigefs_ics.py` - Remove boto initialization (caused failures during testing) and calls to download data from s3 buckets - Output netCDF file is still has `mlgefs` prefix, which is consistent with [exaigefs_prep.sh](https://github.com/NOAA-EMC/aigefs/blob/62f3cb50438e13fce5d95ddd3a09301a68ca1214/scripts/exaigefs_prep.sh#L40) - Both scripts - Remove colon characters from intermediate file names - Explicitly locate files instead of looping and globbing (#36)
This PR:
--debugcan be used to enable verbose loggingUsage:
❯ gen_ics --help usage: gen_ics [-h] {gfs,gefs} ... Download IC data for GFS or GEFS positional arguments: {gfs,gefs} Model to download and process initial conditions for [GFS | GEFS] gfs Download GFS data gefs Download GEFS ensemble data options: -h, --help show this help message and exitDetailed usage for
gfs|gefscan be obtained as:❯ gen_ics gfs --help usage: gen_ics gfs [-h] --current-cycle YYYYMMDDHH [--source {local,s3}] [--target TARGET] [--bucket-name BUCKET_NAME] [--bucket-root-directory BUCKET_ROOT_DIRECTORY] [--comroot COMROOT] [--num-levels NUM_LEVELS] [--varinfo-yaml VARINFO_YAML] [--output OUTPUT] [--debug] [--download-only] options: -h, --help show this help message and exit --current-cycle YYYYMMDDHH Datetime to download and process initial conditions for in YYYYMMDDHH format --source {local,s3} Data source for getting model grib2 data --target TARGET Target directory to store grib2 model data into --bucket-name BUCKET_NAME S3 bucket name. [default: noaa-gfs-bdp-pds (for GFS), noaa-ncepdev-none-ca-ufs- cpldcld (for GEFS)] --bucket-root-directory BUCKET_ROOT_DIRECTORY S3 bucket root directory. [default: (for GFS), Linlin.Cui/gefs_wcoss2 (for GEFS)] --comroot COMROOT Root directory. [default: /lfs/h1/ops/prod/com/gfs/v16.3 (for GFS), /lfs/h1/ops/prod/com/gefs/v12.3 (for GEFS)] --num-levels NUM_LEVELS Number of vertical levels to download from the model data --varinfo-yaml VARINFO_YAML Path to the varinfo YAML file --output OUTPUT Name of the output NetCDF file --debug Set logging level to DEBUG --download-only Only download the data, do not processgefsprovides additional arguments for member:--member {c00,p01,p02,p03,p04,p05,p06,p07,p08,p09,p10,p11,p12,p13,p14,p15,p16,p17,p18,p19,p20,p21,p22,p23,p24,p25,p26,p27,p28,p29,p30} Ensemble memberpytestson github actions are failing due to missing g2c library. I'll need add that to the cache action soon enough.Tests pass locally where pytest is available.