-
Notifications
You must be signed in to change notification settings - Fork 14
Specify files instead of globbing, enable logging, add testing #54
New issue
Have a question about this project? Sign up for a free GitHub account to open an issue and contact its maintainers and the community.
By clicking “Sign up for GitHub”, you agree to our terms of service and privacy statement. We’ll occasionally send you account related emails.
Already on GitHub? Sign in to your account
Closed
aerorahul
wants to merge
22
commits into
NOAA-EMC:production/mlglobal.v1
from
aerorahul:feature/specify_files
Closed
Specify files instead of globbing, enable logging, add testing #54
aerorahul
wants to merge
22
commits into
NOAA-EMC:production/mlglobal.v1
from
aerorahul:feature/specify_files
Conversation
This file contains hidden or bidirectional Unicode text that may be interpreted or compiled differently than what appears below. To review, open the file in an editor that reveals hidden Unicode characters.
Learn more about bidirectional Unicode characters
Collaborator
|
@aerorahul Creating a |
Contributor
Author
|
Installing grib2io on ursa needs g2c. Discussed with @LinlinCui-NOAA and got that resolved. Since this is using |
aerorahul
pushed a commit
that referenced
this pull request
Oct 10, 2025
This PR includes the last changes that SPA team made to - `oper/gen_aigfs_ics.py` - `oper/gen_aigefs_ics.py` to operationalize the scripts. These changes are a band-aid fix that will need a coordinated fix in the next release. A PR such as #54 would likely resolve these issues. Some key areas we will need to improve on include: - `oper/gen_aigefs_ics.py` - Remove boto initialization (caused failures during testing) and calls to download data from s3 buckets - Output netCDF file is still has `mlgefs` prefix, which is consistent with [exaigefs_prep.sh](https://github.com/NOAA-EMC/aigefs/blob/62f3cb50438e13fce5d95ddd3a09301a68ca1214/scripts/exaigefs_prep.sh#L40) - Both scripts - Remove colon characters from intermediate file names - Explicitly locate files instead of looping and globbing (#36)
Sign up for free
to join this conversation on GitHub.
Already have an account?
Sign in to comment
Add this suggestion to a batch that can be applied as a single commit.
This suggestion is invalid because no changes were made to the code.
Suggestions cannot be applied while the pull request is closed.
Suggestions cannot be applied while viewing a subset of changes.
Only one suggestion per line can be applied in a batch.
Add this suggestion to a batch that can be applied as a single commit.
Applying suggestions on deleted lines is not supported.
You must change the existing code in this line in order to create a valid suggestion.
Outdated suggestions cannot be applied.
This suggestion has been applied or marked resolved.
Suggestions cannot be applied from pending reviews.
Suggestions cannot be applied on multi-line comments.
Suggestions cannot be applied while the pull request is queued to merge.
Suggestion cannot be applied right now. Please check back later.
This PR:
--debugcan be used to enable verbose loggingUsage:
❯ gen_ics --help usage: gen_ics [-h] {gfs,gefs} ... Download IC data for GFS or GEFS positional arguments: {gfs,gefs} Model to download and process initial conditions for [GFS | GEFS] gfs Download GFS data gefs Download GEFS ensemble data options: -h, --help show this help message and exitDetailed usage for
gfs|gefscan be obtained as:❯ gen_ics gfs --help usage: gen_ics gfs [-h] --current-cycle YYYYMMDDHH [--source {local,s3}] [--target TARGET] [--bucket-name BUCKET_NAME] [--bucket-root-directory BUCKET_ROOT_DIRECTORY] [--comroot COMROOT] [--num-levels NUM_LEVELS] [--varinfo-yaml VARINFO_YAML] [--output OUTPUT] [--debug] [--download-only] options: -h, --help show this help message and exit --current-cycle YYYYMMDDHH Datetime to download and process initial conditions for in YYYYMMDDHH format --source {local,s3} Data source for getting model grib2 data --target TARGET Target directory to store grib2 model data into --bucket-name BUCKET_NAME S3 bucket name. [default: noaa-gfs-bdp-pds (for GFS), noaa-ncepdev-none-ca-ufs- cpldcld (for GEFS)] --bucket-root-directory BUCKET_ROOT_DIRECTORY S3 bucket root directory. [default: (for GFS), Linlin.Cui/gefs_wcoss2 (for GEFS)] --comroot COMROOT Root directory. [default: /lfs/h1/ops/prod/com/gfs/v16.3 (for GFS), /lfs/h1/ops/prod/com/gefs/v12.3 (for GEFS)] --num-levels NUM_LEVELS Number of vertical levels to download from the model data --varinfo-yaml VARINFO_YAML Path to the varinfo YAML file --output OUTPUT Name of the output NetCDF file --debug Set logging level to DEBUG --download-only Only download the data, do not processgefsprovides additional arguments for member:--member {c00,p01,p02,p03,p04,p05,p06,p07,p08,p09,p10,p11,p12,p13,p14,p15,p16,p17,p18,p19,p20,p21,p22,p23,p24,p25,p26,p27,p28,p29,p30} Ensemble memberpytestson github actions are failing due to missing g2c library. I'll need add that to the cache action soon enough.Tests pass locally where pytest is available.